code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : str = { '''configuration_distilbert''': [ '''DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''DistilBertConfig''', '''DistilBertOnnxConfig''', ], '''tokenization_distilbert''': ['''DistilBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[Any] = ['''DistilBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = [ '''DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''DistilBertForMaskedLM''', '''DistilBertForMultipleChoice''', '''DistilBertForQuestionAnswering''', '''DistilBertForSequenceClassification''', '''DistilBertForTokenClassification''', '''DistilBertModel''', '''DistilBertPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[Any] = [ '''TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFDistilBertForMaskedLM''', '''TFDistilBertForMultipleChoice''', '''TFDistilBertForQuestionAnswering''', '''TFDistilBertForSequenceClassification''', '''TFDistilBertForTokenClassification''', '''TFDistilBertMainLayer''', '''TFDistilBertModel''', '''TFDistilBertPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : List[Any] = [ '''FlaxDistilBertForMaskedLM''', '''FlaxDistilBertForMultipleChoice''', '''FlaxDistilBertForQuestionAnswering''', '''FlaxDistilBertForSequenceClassification''', '''FlaxDistilBertForTokenClassification''', '''FlaxDistilBertModel''', '''FlaxDistilBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
38
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') __lowercase : Dict = logging.getLogger(__name__) @dataclass class __lowercase : lowerCamelCase : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) lowerCamelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) @dataclass class __lowercase : lowerCamelCase : Optional[str] = field(default=_lowercase , metadata={"help": "The input training data file (a text file)."} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={"help": "The number of processes to use for the preprocessing."} , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "The maximum total input sequence length after tokenization. If passed, sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowerCamelCase : bool = field( default=_lowercase , metadata={ "help": ( "Whether to pad all samples to the maximum sentence length. " "If False, will pad the samples dynamically when batching to the maximum length in the batch. More " "efficient on GPU but very bad for TPU." ) } , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def UpperCAmelCase__ (self ): if self.train_file is not None: lowerCamelCase_ : Optional[Any] = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: lowerCamelCase_ : Optional[Any] = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class __lowercase : lowerCamelCase : PreTrainedTokenizerBase lowerCamelCase : Union[bool, str, PaddingStrategy] = True lowerCamelCase : Optional[int] = None lowerCamelCase : Optional[int] = None def __call__(self , A ): lowerCamelCase_ : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' lowerCamelCase_ : str = [feature.pop(A ) for feature in features] lowerCamelCase_ : Any = len(A ) lowerCamelCase_ : List[Any] = len(features[0]['''input_ids'''] ) lowerCamelCase_ : Union[str, Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(A )] for feature in features ] lowerCamelCase_ : str = list(chain(*A ) ) lowerCamelCase_ : Any = self.tokenizer.pad( A , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten lowerCamelCase_ : int = {k: v.view(A , A , -1 ) for k, v in batch.items()} # Add back labels lowerCamelCase_ : Tuple = torch.tensor(A , dtype=torch.intaa ) return batch def lowercase_ ( ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase_ : int = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ : List[str] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , _lowercase , _lowercase ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCamelCase_ : Optional[int] = training_args.get_process_log_level() logger.setLevel(_lowercase ) datasets.utils.logging.set_verbosity(_lowercase ) transformers.utils.logging.set_verbosity(_lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCamelCase_ : Any = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCamelCase_ : str = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: lowerCamelCase_ : Optional[Any] = {} if data_args.train_file is not None: lowerCamelCase_ : Union[str, Any] = data_args.train_file if data_args.validation_file is not None: lowerCamelCase_ : Tuple = data_args.validation_file lowerCamelCase_ : Optional[Any] = data_args.train_file.split('''.''' )[-1] lowerCamelCase_ : Dict = load_dataset( _lowercase , data_files=_lowercase , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. lowerCamelCase_ : Optional[Any] = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase_ : str = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase_ : Any = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase_ : List[Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. lowerCamelCase_ : int = [F"""ending{i}""" for i in range(4 )] lowerCamelCase_ : List[Any] = '''sent1''' lowerCamelCase_ : Dict = '''sent2''' if data_args.max_seq_length is None: lowerCamelCase_ : str = tokenizer.model_max_length if max_seq_length > 1_024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) lowerCamelCase_ : Optional[int] = 1_024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) lowerCamelCase_ : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(_lowercase ): lowerCamelCase_ : Tuple = [[context] * 4 for context in examples[context_name]] lowerCamelCase_ : List[Any] = examples[question_header_name] lowerCamelCase_ : Optional[Any] = [ [F"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(_lowercase ) ] # Flatten out lowerCamelCase_ : Optional[Any] = list(chain(*_lowercase ) ) lowerCamelCase_ : List[Any] = list(chain(*_lowercase ) ) # Tokenize lowerCamelCase_ : List[str] = tokenizer( _lowercase , _lowercase , truncation=_lowercase , max_length=_lowercase , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(_lowercase ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) lowerCamelCase_ : Union[str, Any] = raw_datasets['''train'''] if data_args.max_train_samples is not None: lowerCamelCase_ : List[str] = min(len(_lowercase ) , data_args.max_train_samples ) lowerCamelCase_ : List[str] = train_dataset.select(range(_lowercase ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): lowerCamelCase_ : Dict = train_dataset.map( _lowercase , batched=_lowercase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) lowerCamelCase_ : Optional[int] = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: lowerCamelCase_ : Optional[int] = min(len(_lowercase ) , data_args.max_eval_samples ) lowerCamelCase_ : Any = eval_dataset.select(range(_lowercase ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): lowerCamelCase_ : Tuple = eval_dataset.map( _lowercase , batched=_lowercase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator lowerCamelCase_ : int = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=_lowercase , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(_lowercase ): lowerCamelCase_, lowerCamelCase_ : Optional[Any] = eval_predictions lowerCamelCase_ : Any = np.argmax(_lowercase , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer lowerCamelCase_ : Any = Trainer( model=_lowercase , args=_lowercase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=_lowercase , data_collator=_lowercase , compute_metrics=_lowercase , ) # Training if training_args.do_train: lowerCamelCase_ : int = None if training_args.resume_from_checkpoint is not None: lowerCamelCase_ : Dict = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCamelCase_ : List[Any] = last_checkpoint lowerCamelCase_ : Dict = trainer.train(resume_from_checkpoint=_lowercase ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCamelCase_ : Any = train_result.metrics lowerCamelCase_ : Union[str, Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_lowercase ) ) lowerCamelCase_ : List[Any] = min(_lowercase , len(_lowercase ) ) trainer.log_metrics('''train''' , _lowercase ) trainer.save_metrics('''train''' , _lowercase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCamelCase_ : str = trainer.evaluate() lowerCamelCase_ : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_lowercase ) lowerCamelCase_ : Union[str, Any] = min(_lowercase , len(_lowercase ) ) trainer.log_metrics('''eval''' , _lowercase ) trainer.save_metrics('''eval''' , _lowercase ) lowerCamelCase_ : List[str] = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**_lowercase ) else: trainer.create_model_card(**_lowercase ) def lowercase_ ( _lowercase ) -> Dict: '''simple docstring''' main() if __name__ == "__main__": main()
318
0
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType UpperCAmelCase = None UpperCAmelCase = "<" if sys.byteorder == "little" else ">" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image UpperCAmelCase = [ np.dtype('|b1'), np.dtype('|u1'), np.dtype('<u2'), np.dtype('>u2'), np.dtype('<i2'), np.dtype('>i2'), np.dtype('<u4'), np.dtype('>u4'), np.dtype('<i4'), np.dtype('>i4'), np.dtype('<f4'), np.dtype('>f4'), np.dtype('<f8'), np.dtype('>f8'), ] @dataclass class __snake_case: '''simple docstring''' UpperCAmelCase : int = True UpperCAmelCase : Optional[int] = None # Automatically constructed UpperCAmelCase : Dict = "PIL.Image.Image" UpperCAmelCase : Union[str, Any] = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) UpperCAmelCase : Optional[int] = field(default="Image" , init=_SCREAMING_SNAKE_CASE , repr=_SCREAMING_SNAKE_CASE ) def __call__( self ) -> Dict: return self.pa_type def __snake_case ( self , A_ ) -> dict: if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if isinstance(A_ , A_ ): lowerCAmelCase = np.array(A_ ) if isinstance(A_ , A_ ): return {"path": value, "bytes": None} elif isinstance(A_ , A_ ): return {"path": None, "bytes": value} elif isinstance(A_ , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(A_ ) elif isinstance(A_ , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(A_ ) elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( f'An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def __snake_case ( self , A_ , A_=None ) -> "PIL.Image.Image": if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Image(decode=True) instead.""" ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support decoding images, please install 'Pillow'.""" ) if token_per_repo_id is None: lowerCAmelCase = {} lowerCAmelCase = value["""path"""], value["""bytes"""] if bytes_ is None: if path is None: raise ValueError(f'An image should have one of \'path\' or \'bytes\' but both are None in {value}.' ) else: if is_local_path(A_ ): lowerCAmelCase = PIL.Image.open(A_ ) else: lowerCAmelCase = path.split("""::""" )[-1] try: lowerCAmelCase = string_to_dict(A_ , config.HUB_DATASETS_URL )["""repo_id"""] lowerCAmelCase = token_per_repo_id.get(A_ ) except ValueError: lowerCAmelCase = None with xopen(A_ , """rb""" , use_auth_token=A_ ) as f: lowerCAmelCase = BytesIO(f.read() ) lowerCAmelCase = PIL.Image.open(bytes_ ) else: lowerCAmelCase = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def __snake_case ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return ( self if self.decode else { "bytes": Value("""binary""" ), "path": Value("""string""" ), } ) def __snake_case ( self , A_ ) -> pa.StructArray: if pa.types.is_string(storage.type ): lowerCAmelCase = pa.array([None] * len(A_ ) , type=pa.binary() ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): lowerCAmelCase = pa.array([None] * len(A_ ) , type=pa.string() ) lowerCAmelCase = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: lowerCAmelCase = storage.field("""bytes""" ) else: lowerCAmelCase = pa.array([None] * len(A_ ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: lowerCAmelCase = storage.field("""path""" ) else: lowerCAmelCase = pa.array([None] * len(A_ ) , type=pa.string() ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): lowerCAmelCase = pa.array( [encode_np_array(np.array(A_ ) )["""bytes"""] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) lowerCAmelCase = pa.array([None] * len(A_ ) , type=pa.string() ) lowerCAmelCase = pa.StructArray.from_arrays( [bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(A_ , self.pa_type ) def __snake_case ( self , A_ ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(A_ ): with xopen(A_ , """rb""" ) as f: lowerCAmelCase = f.read() return bytes_ lowerCAmelCase = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) lowerCAmelCase = pa.array( [os.path.basename(A_ ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(A_ , self.pa_type ) def _snake_case ( ) -> str: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() lowerCAmelCase = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _snake_case ( _SCREAMING_SNAKE_CASE : "PIL.Image.Image" ) -> Tuple: """simple docstring""" lowerCAmelCase = BytesIO() if image.format in list_image_compression_formats(): lowerCAmelCase = image.format else: lowerCAmelCase = """PNG""" if image.mode in ["""1""", """L""", """LA""", """RGB""", """RGBA"""] else """TIFF""" image.save(A__ , format=A__ ) return buffer.getvalue() def _snake_case ( _SCREAMING_SNAKE_CASE : "PIL.Image.Image" ) -> List[Any]: """simple docstring""" if hasattr(A__ , """filename""" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(A__ )} def _snake_case ( _SCREAMING_SNAKE_CASE : np.ndarray ) -> List[Any]: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) lowerCAmelCase = array.dtype lowerCAmelCase = dtype.byteorder if dtype.byteorder != """=""" else _NATIVE_BYTEORDER lowerCAmelCase = dtype.kind lowerCAmelCase = dtype.itemsize lowerCAmelCase = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: lowerCAmelCase = np.dtype("""|u1""" ) if dtype_kind not in ["u", "i"]: raise TypeError( f'Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.' ) if dtype is not dest_dtype: warnings.warn(f'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: lowerCAmelCase = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: lowerCAmelCase = dtype_byteorder + dtype_kind + str(A__ ) lowerCAmelCase = np.dtype(A__ ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(f'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( f'Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}' ) lowerCAmelCase = PIL.Image.fromarray(array.astype(A__ ) ) return {"path": None, "bytes": image_to_bytes(A__ )} def _snake_case ( _SCREAMING_SNAKE_CASE : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ) -> int: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if objs: lowerCAmelCase = first_non_null_value(A__ ) if isinstance(A__ , A__ ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(A__ , np.ndarray ): lowerCAmelCase = no_op_if_value_is_null(A__ ) return [obj_to_image_dict_func(A__ ) for obj in objs] elif isinstance(A__ , PIL.Image.Image ): lowerCAmelCase = no_op_if_value_is_null(A__ ) return [obj_to_image_dict_func(A__ ) for obj in objs] else: return objs else: return objs
351
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = '▁' UpperCAmelCase = { 'vocab_file': 'vocab.json', 'spm_file': 'sentencepiece.bpe.model', 'tokenizer_config_file': 'tokenizer_config.json', } UpperCAmelCase = { 'vocab_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json', }, 'spm_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_config_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json', }, } UpperCAmelCase = { 'facebook/m2m100_418M': 1024, } # fmt: off UpperCAmelCase = { 'm2m100': ['af', 'am', 'ar', 'ast', 'az', 'ba', 'be', 'bg', 'bn', 'br', 'bs', 'ca', 'ceb', 'cs', 'cy', 'da', 'de', 'el', 'en', 'es', 'et', 'fa', 'ff', 'fi', 'fr', 'fy', 'ga', 'gd', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hy', 'id', 'ig', 'ilo', 'is', 'it', 'ja', 'jv', 'ka', 'kk', 'km', 'kn', 'ko', 'lb', 'lg', 'ln', 'lo', 'lt', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'my', 'ne', 'nl', 'no', 'ns', 'oc', 'or', 'pa', 'pl', 'ps', 'pt', 'ro', 'ru', 'sd', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'ss', 'su', 'sv', 'sw', 'ta', 'th', 'tl', 'tn', 'tr', 'uk', 'ur', 'uz', 'vi', 'wo', 'xh', 'yi', 'yo', 'zh', 'zu'], 'wmt21': ['en', 'ha', 'is', 'ja', 'cs', 'ru', 'zh', 'de'] } class __snake_case( _lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Optional[int] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : int = ["input_ids", "attention_mask"] UpperCAmelCase : List[int] = [] UpperCAmelCase : List[int] = [] def __init__( self , A_ , A_ , A_=None , A_=None , A_="<s>" , A_="</s>" , A_="</s>" , A_="<pad>" , A_="<unk>" , A_="m2m100" , A_ = None , A_=8 , **A_ , ) -> None: lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs lowerCAmelCase = language_codes lowerCAmelCase = FAIRSEQ_LANGUAGE_CODES[language_codes] lowerCAmelCase = {lang_code: f'__{lang_code}__' for lang_code in fairseq_language_code} lowerCAmelCase = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(A_ ) for lang_code in fairseq_language_code if self.get_lang_token(A_ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=A_ , tgt_lang=A_ , bos_token=A_ , eos_token=A_ , sep_token=A_ , unk_token=A_ , pad_token=A_ , language_codes=A_ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=A_ , **A_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = load_json(A_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} lowerCAmelCase = spm_file lowerCAmelCase = load_spm(A_ , self.sp_model_kwargs ) lowerCAmelCase = len(self.encoder ) lowerCAmelCase = { self.get_lang_token(A_ ): self.encoder_size + i for i, lang_code in enumerate(A_ ) } lowerCAmelCase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(A_ )} lowerCAmelCase = {v: k for k, v in self.lang_token_to_id.items()} lowerCAmelCase = src_lang if src_lang is not None else """en""" lowerCAmelCase = tgt_lang lowerCAmelCase = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) lowerCAmelCase = num_madeup_words @property def __snake_case ( self ) -> int: return len(self.encoder ) + len(self.lang_token_to_id ) @property def __snake_case ( self ) -> str: return self._src_lang @src_lang.setter def __snake_case ( self , A_ ) -> None: lowerCAmelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __snake_case ( self , A_ ) -> List[str]: return self.sp_model.encode(A_ , out_type=A_ ) def __snake_case ( self , A_ ) -> Any: if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(A_ , self.encoder[self.unk_token] ) def __snake_case ( self , A_ ) -> str: if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(A_ , self.unk_token ) def __snake_case ( self , A_ ) -> List[str]: lowerCAmelCase = [] lowerCAmelCase = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A_ ) + token lowerCAmelCase = [] else: current_sub_tokens.append(A_ ) out_string += self.sp_model.decode(A_ ) return out_string.strip() def __snake_case ( self , A_ , A_ = None , A_ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) lowerCAmelCase = [1] * len(self.prefix_tokens ) lowerCAmelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(A_ )) + suffix_ones return prefix_ones + ([0] * len(A_ )) + ([0] * len(A_ )) + suffix_ones def __snake_case ( self , A_ , A_ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __snake_case ( self ) -> Dict: lowerCAmelCase = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Dict: lowerCAmelCase = self.__dict__.copy() lowerCAmelCase = None return state def __setstate__( self , A_ ) -> None: lowerCAmelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCAmelCase = {} lowerCAmelCase = load_spm(self.spm_file , self.sp_model_kwargs ) def __snake_case ( self , A_ , A_ = None ) -> Tuple[str]: lowerCAmelCase = Path(A_ ) if not save_dir.is_dir(): raise OSError(f'{save_directory} should be a directory' ) lowerCAmelCase = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""vocab_file"""] ) lowerCAmelCase = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""spm_file"""] ) save_json(self.encoder , A_ ) if os.path.abspath(self.spm_file ) != os.path.abspath(A_ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , A_ ) elif not os.path.isfile(self.spm_file ): with open(A_ , """wb""" ) as fi: lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(A_ ) return (str(A_ ), str(A_ )) def __snake_case ( self , A_ , A_ = "en" , A_ = None , A_ = "ro" , **A_ , ) -> BatchEncoding: lowerCAmelCase = src_lang lowerCAmelCase = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(A_ , A_ , **A_ ) def __snake_case ( self , A_ , A_ , A_ , **A_ ) -> str: if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) lowerCAmelCase = src_lang lowerCAmelCase = self(A_ , add_special_tokens=A_ , **A_ ) lowerCAmelCase = self.get_lang_id(A_ ) lowerCAmelCase = tgt_lang_id return inputs def __snake_case ( self ) -> Any: self.set_src_lang_special_tokens(self.src_lang ) def __snake_case ( self ) -> Optional[int]: self.set_tgt_lang_special_tokens(self.tgt_lang ) def __snake_case ( self , A_ ) -> None: lowerCAmelCase = self.get_lang_token(A_ ) lowerCAmelCase = self.lang_token_to_id[lang_token] lowerCAmelCase = [self.cur_lang_id] lowerCAmelCase = [self.eos_token_id] def __snake_case ( self , A_ ) -> None: lowerCAmelCase = self.get_lang_token(A_ ) lowerCAmelCase = self.lang_token_to_id[lang_token] lowerCAmelCase = [self.cur_lang_id] lowerCAmelCase = [self.eos_token_id] def __snake_case ( self , A_ ) -> str: return self.lang_code_to_token[lang] def __snake_case ( self , A_ ) -> int: lowerCAmelCase = self.get_lang_token(A_ ) return self.lang_token_to_id[lang_token] def _snake_case ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Dict[str, Any] ) -> sentencepiece.SentencePieceProcessor: """simple docstring""" lowerCAmelCase = sentencepiece.SentencePieceProcessor(**_SCREAMING_SNAKE_CASE ) spm.Load(str(_SCREAMING_SNAKE_CASE ) ) return spm def _snake_case ( _SCREAMING_SNAKE_CASE : str ) -> Union[Dict, List]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: return json.load(_SCREAMING_SNAKE_CASE ) def _snake_case ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : str ) -> None: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , indent=2 )
187
0
'''simple docstring''' def __snake_case( _lowerCAmelCase = 1_000_000 ) -> int: snake_case__ : Dict = set(range(3 , _lowerCAmelCase , 2 ) ) primes.add(2 ) for p in range(3 , _lowerCAmelCase , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , _lowerCAmelCase , _lowerCAmelCase ) ) ) snake_case__ : Optional[int] = [float(_lowerCAmelCase ) for n in range(limit + 1 )] for p in primes: for n in range(_lowerCAmelCase , limit + 1 , _lowerCAmelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"{solution() = }")
35
'''simple docstring''' # Function to print upper half of diamond (pyramid) def __snake_case( _lowerCAmelCase ) -> Any: for i in range(0 , _lowerCAmelCase ): for _ in range(0 , n - i - 1 ): # printing spaces print(""" """ , end="""""" ) for _ in range(0 , i + 1 ): # printing stars print("""* """ , end="""""" ) print() def __snake_case( _lowerCAmelCase ) -> List[str]: for i in range(_lowerCAmelCase , 0 , -1 ): for _ in range(_lowerCAmelCase , 0 , -1 ): # printing stars print("""* """ , end="""""" ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(""" """ , end="""""" ) def __snake_case( _lowerCAmelCase ) -> List[Any]: if n <= 0: print(""" ... .... nothing printing :(""" ) return floyd(_lowerCAmelCase ) # upper half reverse_floyd(_lowerCAmelCase ) # lower half if __name__ == "__main__": print(R"| /\ | |- | |- |--| |\ /| |-") print(R"|/ \| |- |_ |_ |__| | \/ | |_") __a = 1 while K: __a = int(input("enter the number and , and see the magic : ")) print() pretty_print(user_number) __a = int(input("press 0 to exit... and 1 to continue...")) print("Good Bye...")
35
1
"""simple docstring""" def _snake_case ( _snake_case : str ) -> bool: '''simple docstring''' if not all(x.isalpha() for x in string ): raise ValueError('String must only contain alphabetic characters.' ) _A = sorted(string.lower() ) return len(_snake_case ) == len(set(_snake_case ) ) if __name__ == "__main__": a = input('''Enter a string ''').strip() a = is_isogram(input_str) print(F'''{input_str} is {"an" if isogram else "not an"} isogram.''')
271
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = (IPNDMScheduler,) UpperCAmelCase : Optional[Any] = (('''num_inference_steps''', 50),) def lowerCAmelCase_ ( self : Union[str, Any] , **_UpperCAmelCase : List[Any] ): _A = {'num_train_timesteps': 1_000} config.update(**_UpperCAmelCase ) return config def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : Optional[int]=0 , **_UpperCAmelCase : Union[str, Any] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] if time_step is None: _A = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : str ): pass def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : Any=0 , **_UpperCAmelCase : Any ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals (must be after setting timesteps) _A = dummy_past_residuals[:] if time_step is None: _A = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residual (must be after setting timesteps) _A = dummy_past_residuals[:] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : List[str] , **_UpperCAmelCase : Optional[int] ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) _A = 10 _A = self.dummy_model() _A = self.dummy_sample_deter scheduler.set_timesteps(_UpperCAmelCase ) for i, t in enumerate(scheduler.timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample for i, t in enumerate(scheduler.timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample return sample def lowerCAmelCase_ ( self : Union[str, Any] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample if num_inference_steps is not None and hasattr(_UpperCAmelCase , 'set_timesteps' ): scheduler.set_timesteps(_UpperCAmelCase ) elif num_inference_steps is not None and not hasattr(_UpperCAmelCase , 'set_timesteps' ): _A = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] _A = dummy_past_residuals[:] _A = scheduler.timesteps[5] _A = scheduler.timesteps[6] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowerCAmelCase_ ( self : Tuple ): for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCAmelCase , time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=_UpperCAmelCase , time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] ): _A = self.full_loop() _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_mean.item() - 2_540_529 ) < 10
271
1
'''simple docstring''' from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets lowerCamelCase__ = "\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n" lowerCamelCase__ = "\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n" lowerCamelCase__ = "\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n \"accuracy\": Accuracy\n \"f1\": F1 score\n \"pearson\": Pearson Correlation\n \"spearmanr\": Spearman Correlation\n \"matthews_correlation\": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'stsb')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)})\n {'pearson': 1.0, 'spearmanr': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'cola')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return float((preds == labels).mean() ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = simple_accuracy(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) _UpperCAmelCase : List[str] = float(fa_score(y_true=__SCREAMING_SNAKE_CASE , y_pred=__SCREAMING_SNAKE_CASE ) ) return { "accuracy": acc, "f1": fa, } def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[int] = float(pearsonr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )[0] ) _UpperCAmelCase : Tuple = float(spearmanr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : str ) ->Tuple: '''simple docstring''' if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), "references": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : int , lowerCamelCase__ : Optional[Any] ) ->List[Any]: '''simple docstring''' if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )} elif self.config_name == "stsb": return pearson_and_spearman(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )} else: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" )
234
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _lowercase : Any = (7_2_0, 1_2_8_0) # Height, Width _lowercase : List[Any] = (0.4, 0.6) # if height or width lower than this scale, drop it. _lowercase : str = 1 / 1_0_0 _lowercase : Any = "" _lowercase : Union[str, Any] = "" _lowercase : Optional[int] = "" _lowercase : List[Any] = 2_5_0 def snake_case_ ( ): """simple docstring""" lowercase_ , lowercase_ : Any = get_dataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for index in range(__SCREAMING_SNAKE_CASE ): lowercase_ : str = random.sample(range(len(__SCREAMING_SNAKE_CASE ) ) , 4 ) lowercase_ , lowercase_ , lowercase_ : Any = update_image_and_anno( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , filter_scale=__SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' lowercase_ : int = random_chars(32 ) lowercase_ : str = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] lowercase_ : int = F'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(F'''{file_root}.jpg''' , __SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) lowercase_ : List[Any] = [] for anno in new_annos: lowercase_ : List[Any] = anno[3] - anno[1] lowercase_ : List[str] = anno[4] - anno[2] lowercase_ : Dict = anno[1] + width / 2 lowercase_ : Dict = anno[2] + height / 2 lowercase_ : int = F'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(__SCREAMING_SNAKE_CASE ) with open(F'''{file_root}.txt''' , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def snake_case_ ( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str ): """simple docstring""" lowercase_ : Optional[Any] = [] lowercase_ : Optional[Any] = [] for label_file in glob.glob(os.path.join(__SCREAMING_SNAKE_CASE , '''*.txt''' ) ): lowercase_ : int = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(__SCREAMING_SNAKE_CASE ) as in_file: lowercase_ : List[str] = in_file.readlines() lowercase_ : Optional[Any] = os.path.join(__SCREAMING_SNAKE_CASE , F'''{label_name}.jpg''' ) lowercase_ : Optional[int] = [] for obj_list in obj_lists: lowercase_ : List[str] = obj_list.rstrip('''\n''' ).split(''' ''' ) lowercase_ : Optional[int] = float(obj[1] ) - float(obj[3] ) / 2 lowercase_ : Any = float(obj[2] ) - float(obj[4] ) / 2 lowercase_ : str = float(obj[1] ) + float(obj[3] ) / 2 lowercase_ : List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(__SCREAMING_SNAKE_CASE ) labels.append(__SCREAMING_SNAKE_CASE ) return img_paths, labels def snake_case_ ( __SCREAMING_SNAKE_CASE : list , __SCREAMING_SNAKE_CASE : list , __SCREAMING_SNAKE_CASE : list[int] , __SCREAMING_SNAKE_CASE : tuple[int, int] , __SCREAMING_SNAKE_CASE : tuple[float, float] , __SCREAMING_SNAKE_CASE : float = 0.0 , ): """simple docstring""" lowercase_ : List[Any] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) lowercase_ : Tuple = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) lowercase_ : List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) lowercase_ : Optional[int] = int(scale_x * output_size[1] ) lowercase_ : Dict = int(scale_y * output_size[0] ) lowercase_ : Union[str, Any] = [] lowercase_ : List[Any] = [] for i, index in enumerate(__SCREAMING_SNAKE_CASE ): lowercase_ : Union[str, Any] = all_img_list[index] path_list.append(__SCREAMING_SNAKE_CASE ) lowercase_ : int = all_annos[index] lowercase_ : Dict = cva.imread(__SCREAMING_SNAKE_CASE ) if i == 0: # top-left lowercase_ : Optional[Any] = cva.resize(__SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) lowercase_ : Tuple = img for bbox in img_annos: lowercase_ : Optional[int] = bbox[1] * scale_x lowercase_ : Optional[Any] = bbox[2] * scale_y lowercase_ : str = bbox[3] * scale_x lowercase_ : Tuple = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right lowercase_ : Dict = cva.resize(__SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) lowercase_ : Dict = img for bbox in img_annos: lowercase_ : int = scale_x + bbox[1] * (1 - scale_x) lowercase_ : Dict = bbox[2] * scale_y lowercase_ : Optional[int] = scale_x + bbox[3] * (1 - scale_x) lowercase_ : int = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left lowercase_ : List[Any] = cva.resize(__SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) lowercase_ : List[str] = img for bbox in img_annos: lowercase_ : Any = bbox[1] * scale_x lowercase_ : Optional[int] = scale_y + bbox[2] * (1 - scale_y) lowercase_ : str = bbox[3] * scale_x lowercase_ : Optional[int] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right lowercase_ : int = cva.resize( __SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) lowercase_ : List[str] = img for bbox in img_annos: lowercase_ : int = scale_x + bbox[1] * (1 - scale_x) lowercase_ : Any = scale_y + bbox[2] * (1 - scale_y) lowercase_ : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) lowercase_ : int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: lowercase_ : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def snake_case_ ( __SCREAMING_SNAKE_CASE : int ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" lowercase_ : Any = ascii_lowercase + digits return "".join(random.choice(__SCREAMING_SNAKE_CASE ) for _ in range(__SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print("DONE ✅")
93
0
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING UpperCAmelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase_ ) class a ( lowerCAmelCase_ ): def __init__( self : Union[str, Any] , *__lowerCAmelCase : Optional[Any] , **__lowerCAmelCase : int ): super().__init__(*__lowerCAmelCase , **__lowerCAmelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == """tf""" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def lowerCAmelCase_ ( self : int , __lowerCAmelCase : Dict=None , __lowerCAmelCase : List[Any]=None , __lowerCAmelCase : Tuple=None ): _UpperCAmelCase = {} _UpperCAmelCase = {} if prompt is not None: _UpperCAmelCase = prompt if generate_kwargs is not None: _UpperCAmelCase = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: _UpperCAmelCase = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( """'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter,""" """ please use only one""" ) _UpperCAmelCase = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self : Optional[Any] , __lowerCAmelCase : Union[str, List[str], "Image.Image", List["Image.Image"]] , **__lowerCAmelCase : Optional[int] ): return super().__call__(__lowerCAmelCase , **__lowerCAmelCase ) def lowerCAmelCase_ ( self : List[str] , __lowerCAmelCase : int , __lowerCAmelCase : List[str]=None ): _UpperCAmelCase = load_image(__lowerCAmelCase ) if prompt is not None: if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise ValueError( f'''Received an invalid text input, got - {type(__lowerCAmelCase )} - but expected a single string. ''' """Note also that one single text can be provided for conditional image to text generation.""" ) _UpperCAmelCase = self.model.config.model_type if model_type == "git": _UpperCAmelCase = self.image_processor(images=__lowerCAmelCase , return_tensors=self.framework ) _UpperCAmelCase = self.tokenizer(text=__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ).input_ids _UpperCAmelCase = [self.tokenizer.cls_token_id] + input_ids _UpperCAmelCase = torch.tensor(__lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"""input_ids""": input_ids} ) elif model_type == "pix2struct": _UpperCAmelCase = self.image_processor(images=__lowerCAmelCase , header_text=__lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation _UpperCAmelCase = self.image_processor(images=__lowerCAmelCase , return_tensors=self.framework ) _UpperCAmelCase = self.tokenizer(__lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(__lowerCAmelCase ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: _UpperCAmelCase = self.image_processor(images=__lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: _UpperCAmelCase = None return model_inputs def lowerCAmelCase_ ( self : int , __lowerCAmelCase : Tuple , __lowerCAmelCase : List[Any]=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["""input_ids"""] , __lowerCAmelCase ) and all(x is None for x in model_inputs["""input_ids"""] ) ): _UpperCAmelCase = None if generate_kwargs is None: _UpperCAmelCase = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. _UpperCAmelCase = model_inputs.pop(self.model.main_input_name ) _UpperCAmelCase = self.model.generate(__lowerCAmelCase , **__lowerCAmelCase , **__lowerCAmelCase ) return model_outputs def lowerCAmelCase_ ( self : Tuple , __lowerCAmelCase : int ): _UpperCAmelCase = [] for output_ids in model_outputs: _UpperCAmelCase = { """generated_text""": self.tokenizer.decode( __lowerCAmelCase , skip_special_tokens=__lowerCAmelCase , ) } records.append(__lowerCAmelCase ) return records
371
"""simple docstring""" from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class a ( lowerCAmelCase_ ): def __init__( self : Optional[int] , __lowerCAmelCase : Any=None , __lowerCAmelCase : Any=None , *__lowerCAmelCase : Union[str, Any] , **__lowerCAmelCase : Optional[int] ): super().__init__(*__lowerCAmelCase , **__lowerCAmelCase ) if config is None: assert isinstance(self.model , __lowerCAmelCase ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f''' {self.model.__class__}''' ) _UpperCAmelCase = self.model.config else: _UpperCAmelCase = config _UpperCAmelCase = data_args _UpperCAmelCase = self.config.tgt_vocab_size if isinstance(self.config , __lowerCAmelCase ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' """ padding..""" ) if self.args.label_smoothing == 0: _UpperCAmelCase = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _UpperCAmelCase = label_smoothed_nll_loss def lowerCAmelCase_ ( self : Dict , __lowerCAmelCase : int ): if self.optimizer is None: _UpperCAmelCase = ["""bias""", """LayerNorm.weight"""] _UpperCAmelCase = [ { """params""": [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], """weight_decay""": self.args.weight_decay, }, { """params""": [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] _UpperCAmelCase = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _UpperCAmelCase = Adafactor _UpperCAmelCase = {"""scale_parameter""": False, """relative_step""": False} else: _UpperCAmelCase = AdamW _UpperCAmelCase = { """betas""": (self.args.adam_betaa, self.args.adam_betaa), """eps""": self.args.adam_epsilon, } _UpperCAmelCase = self.args.learning_rate if self.sharded_ddp: _UpperCAmelCase = OSS( params=__lowerCAmelCase , optim=__lowerCAmelCase , **__lowerCAmelCase , ) else: _UpperCAmelCase = optimizer_cls(__lowerCAmelCase , **__lowerCAmelCase ) if self.lr_scheduler is None: _UpperCAmelCase = self._get_lr_scheduler(__lowerCAmelCase ) else: # ignoring --lr_scheduler logger.warning("""scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.""" ) def lowerCAmelCase_ ( self : Optional[Any] , __lowerCAmelCase : List[str] ): _UpperCAmelCase = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _UpperCAmelCase = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _UpperCAmelCase = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: _UpperCAmelCase = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__lowerCAmelCase ) return scheduler def lowerCAmelCase_ ( self : Optional[int] ): if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def lowerCAmelCase_ ( self : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple ): if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _UpperCAmelCase = model(**__lowerCAmelCase , use_cache=__lowerCAmelCase )[0] _UpperCAmelCase = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models _UpperCAmelCase , _UpperCAmelCase = model(**__lowerCAmelCase , labels=__lowerCAmelCase , use_cache=__lowerCAmelCase )[:2] else: # compute label smoothed loss _UpperCAmelCase = model(**__lowerCAmelCase , use_cache=__lowerCAmelCase )[0] _UpperCAmelCase = torch.nn.functional.log_softmax(__lowerCAmelCase , dim=-1 ) _UpperCAmelCase , _UpperCAmelCase = self.loss_fn(__lowerCAmelCase , __lowerCAmelCase , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def lowerCAmelCase_ ( self : Optional[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : int ): _UpperCAmelCase = inputs.pop("""labels""" ) _UpperCAmelCase , _UpperCAmelCase = self._compute_loss(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) return loss def lowerCAmelCase_ ( self : List[Any] , __lowerCAmelCase : nn.Module , __lowerCAmelCase : Dict[str, Union[torch.Tensor, Any]] , __lowerCAmelCase : bool , __lowerCAmelCase : Optional[List[str]] = None , ): _UpperCAmelCase = self._prepare_inputs(__lowerCAmelCase ) _UpperCAmelCase = { """max_length""": self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, """num_beams""": self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _UpperCAmelCase = self.model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , **__lowerCAmelCase , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _UpperCAmelCase = self._pad_tensors_to_max_len(__lowerCAmelCase , gen_kwargs["""max_length"""] ) _UpperCAmelCase = inputs.pop("""labels""" ) with torch.no_grad(): # compute loss on predict data _UpperCAmelCase , _UpperCAmelCase = self._compute_loss(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _UpperCAmelCase = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _UpperCAmelCase = self._pad_tensors_to_max_len(__lowerCAmelCase , gen_kwargs["""max_length"""] ) return (loss, logits, labels) def lowerCAmelCase_ ( self : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str] ): # If PAD token is not defined at least EOS token has to be defined _UpperCAmelCase = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( """Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be""" f''' padded to `max_length`={max_length}''' ) _UpperCAmelCase = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) _UpperCAmelCase = tensor return padded_tensor
30
0
'''simple docstring''' def __lowercase ( __lowercase , __lowercase , __lowercase=False ) -> Union[str, Any]: '''simple docstring''' if isinstance(__lowercase , __lowercase ) and isinstance(__lowercase , __lowercase ): _A = len(set_a.intersection(__lowercase ) ) if alternative_union: _A = len(__lowercase ) + len(__lowercase ) else: _A = len(set_a.union(__lowercase ) ) return intersection / union if isinstance(__lowercase , (list, tuple) ) and isinstance(__lowercase , (list, tuple) ): _A = [element for element in set_a if element in set_b] if alternative_union: _A = len(__lowercase ) + len(__lowercase ) return len(__lowercase ) / union else: _A = set_a + [element for element in set_b if element not in set_a] return len(__lowercase ) / len(__lowercase ) return len(__lowercase ) / len(__lowercase ) return None if __name__ == "__main__": lowerCamelCase_ = {'''a''', '''b''', '''c''', '''d''', '''e'''} lowerCamelCase_ = {'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
79
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = StableDiffusionInpaintPipeline lowerCamelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowerCamelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCamelCase__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowerCamelCase__ = frozenset([] ) def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowercase , ) _lowerCamelCase : Optional[Any] = PNDMScheduler(skip_prk_steps=lowercase ) torch.manual_seed(0 ) _lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _lowerCamelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) _lowerCamelCase : int = CLIPTextModel(lowercase ) _lowerCamelCase : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCamelCase : Optional[int] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self , lowercase , lowercase=0 ): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched _lowerCamelCase : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) _lowerCamelCase : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase : Optional[int] = Image.fromarray(np.uinta(lowercase ) ).convert('RGB' ).resize((64, 64) ) _lowerCamelCase : Optional[Any] = Image.fromarray(np.uinta(image + 4 ) ).convert('RGB' ).resize((64, 64) ) if str(lowercase ).startswith('mps' ): _lowerCamelCase : str = torch.manual_seed(lowercase ) else: _lowerCamelCase : int = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : str = { 'prompt': 'A painting of a squirrel eating a burger', 'image': init_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase : Union[str, Any] = self.get_dummy_components() _lowerCamelCase : Union[str, Any] = StableDiffusionInpaintPipeline(**lowercase ) _lowerCamelCase : Tuple = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Dict = self.get_dummy_inputs(lowercase ) _lowerCamelCase : Any = sd_pipe(**lowercase ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _lowerCamelCase : List[Any] = np.array([0.47_27, 0.57_35, 0.39_41, 0.54_46, 0.59_26, 0.43_94, 0.50_62, 0.46_54, 0.44_76] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) _lowerCamelCase : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) _lowerCamelCase : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench.npy' ) _lowerCamelCase : List[Any] = 'stabilityai/stable-diffusion-2-inpainting' _lowerCamelCase : List[str] = StableDiffusionInpaintPipeline.from_pretrained(lowercase , safety_checker=lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : Optional[int] = 'Face of a yellow cat, high resolution, sitting on a park bench' _lowerCamelCase : Tuple = torch.manual_seed(0 ) _lowerCamelCase : Optional[int] = pipe( prompt=lowercase , image=lowercase , mask_image=lowercase , generator=lowercase , output_type='np' , ) _lowerCamelCase : Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def A_ ( self ): _lowerCamelCase : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) _lowerCamelCase : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) _lowerCamelCase : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench_fp16.npy' ) _lowerCamelCase : int = 'stabilityai/stable-diffusion-2-inpainting' _lowerCamelCase : List[Any] = StableDiffusionInpaintPipeline.from_pretrained( lowercase , torch_dtype=torch.floataa , safety_checker=lowercase , ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : List[str] = 'Face of a yellow cat, high resolution, sitting on a park bench' _lowerCamelCase : str = torch.manual_seed(0 ) _lowerCamelCase : List[str] = pipe( prompt=lowercase , image=lowercase , mask_image=lowercase , generator=lowercase , output_type='np' , ) _lowerCamelCase : Tuple = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def A_ ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _lowerCamelCase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) _lowerCamelCase : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) _lowerCamelCase : Optional[Any] = 'stabilityai/stable-diffusion-2-inpainting' _lowerCamelCase : List[Any] = PNDMScheduler.from_pretrained(lowercase , subfolder='scheduler' ) _lowerCamelCase : Union[str, Any] = StableDiffusionInpaintPipeline.from_pretrained( lowercase , safety_checker=lowercase , scheduler=lowercase , torch_dtype=torch.floataa , ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _lowerCamelCase : str = 'Face of a yellow cat, high resolution, sitting on a park bench' _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : Dict = pipe( prompt=lowercase , image=lowercase , mask_image=lowercase , generator=lowercase , num_inference_steps=2 , output_type='np' , ) _lowerCamelCase : int = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
12
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowercase__ = False class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _lowerCamelCase : Dict = torch.manual_seed(0 ) _lowerCamelCase : Dict = pipe( image=lowercase , generator=lowercase , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' , ).images _lowerCamelCase : str = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : Any = np.array([0.04_41, 0.04_69, 0.05_07, 0.05_75, 0.06_32, 0.06_50, 0.08_65, 0.09_09, 0.09_45] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
12
1
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __lowerCamelCase = logging.get_logger(__name__) __lowerCamelCase = {"tokenizer_file": "tokenizer.json"} __lowerCamelCase = { "tokenizer_file": { "bigscience/tokenizer": "https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json", "bigscience/bloom-560m": "https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json", "bigscience/bloom-1b1": "https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json", "bigscience/bloom-1b7": "https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json", "bigscience/bloom-3b": "https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json", "bigscience/bloom-7b1": "https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json", "bigscience/bloom": "https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json", }, } class A__ ( SCREAMING_SNAKE_CASE__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = ["input_ids", "attention_mask"] lowercase = None def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__="<unk>" , UpperCamelCase__="<s>" , UpperCamelCase__="</s>" , UpperCamelCase__="<pad>" , UpperCamelCase__=False , UpperCamelCase__=False , **UpperCamelCase__ , ) -> Optional[Any]: '''simple docstring''' super().__init__( UpperCamelCase__ , UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , unk_token=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ , **UpperCamelCase__ , ) A_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCamelCase__ ) != add_prefix_space: A_ = getattr(UpperCamelCase__ , pre_tok_state.pop("""type""" ) ) A_ = add_prefix_space A_ = pre_tok_class(**UpperCamelCase__ ) A_ = add_prefix_space def snake_case_ ( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> BatchEncoding: '''simple docstring''' A_ = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with''' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def snake_case_ ( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> BatchEncoding: '''simple docstring''' A_ = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with''' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' A_ = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ ) -> List[int]: '''simple docstring''' A_ = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) + [self.eos_token_id] ) if len(UpperCamelCase__ ) > self.model_max_length: A_ = input_ids[-self.model_max_length :] return input_ids
162
from ...processing_utils import ProcessorMixin class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = ['''image_processor''', '''feature_extractor'''] lowerCamelCase__ = '''TvltImageProcessor''' lowerCamelCase__ = '''TvltFeatureExtractor''' def __init__( self : List[str] , __magic_name__ : Any , __magic_name__ : Any ) -> int: super().__init__(image_processor=__magic_name__ , feature_extractor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = image_processor SCREAMING_SNAKE_CASE_ = feature_extractor def __call__( self : List[str] , __magic_name__ : Optional[Any]=None , __magic_name__ : Union[str, Any]=None , __magic_name__ : int=None , __magic_name__ : str=None , __magic_name__ : Any=False , __magic_name__ : int=False , *__magic_name__ : int , **__magic_name__ : Any , ) -> List[Any]: if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) SCREAMING_SNAKE_CASE_ = None if images is not None: SCREAMING_SNAKE_CASE_ = self.image_processor(__magic_name__ , mask_pixel=__magic_name__ , *__magic_name__ , **__magic_name__ ) if images_mixed is not None: SCREAMING_SNAKE_CASE_ = self.image_processor(__magic_name__ , is_mixed=__magic_name__ , *__magic_name__ , **__magic_name__ ) if audio is not None: SCREAMING_SNAKE_CASE_ = self.feature_extractor( __magic_name__ , *__magic_name__ , sampling_rate=__magic_name__ , mask_audio=__magic_name__ , **__magic_name__ ) SCREAMING_SNAKE_CASE_ = {} if audio is not None: output_dict.update(__magic_name__ ) if images is not None: output_dict.update(__magic_name__ ) if images_mixed_dict is not None: output_dict.update(__magic_name__ ) return output_dict @property def __A ( self : Optional[Any] ) -> int: SCREAMING_SNAKE_CASE_ = self.image_processor.model_input_names SCREAMING_SNAKE_CASE_ = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
118
0
"""simple docstring""" import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser __A = logging.getLogger(__name__) torch.set_grad_enabled(False) __A = 'cuda' if torch.cuda.is_available() else 'cpu' def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase=100 , __UpperCamelCase=" " ) -> List[str]: _lowerCAmelCase =text.split(__UpperCamelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__UpperCamelCase ) , __UpperCamelCase )] def _lowerCamelCase(__UpperCamelCase ) -> dict: _lowerCAmelCase , _lowerCAmelCase =[], [] for title, text in zip(documents["""title"""] , documents["""text"""] ): if text is not None: for passage in split_text(__UpperCamelCase ): titles.append(title if title is not None else """""" ) texts.append(__UpperCamelCase ) return {"title": titles, "text": texts} def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> dict: _lowerCAmelCase =ctx_tokenizer( documents["""title"""] , documents["""text"""] , truncation=__UpperCamelCase , padding="""longest""" , return_tensors="""pt""" )["""input_ids"""] _lowerCAmelCase =ctx_encoder(input_ids.to(device=__UpperCamelCase ) , return_dict=__UpperCamelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ) -> int: ###################################### logger.info("""Step 1 - Create the dataset""" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _lowerCAmelCase =load_dataset( """csv""" , data_files=[rag_example_args.csv_path] , split="""train""" , delimiter="""\t""" , column_names=["""title""", """text"""] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _lowerCAmelCase =dataset.map(__UpperCamelCase , batched=__UpperCamelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _lowerCAmelCase =DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__UpperCamelCase ) _lowerCAmelCase =DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _lowerCAmelCase =Features( {"""text""": Value("""string""" ), """title""": Value("""string""" ), """embeddings""": Sequence(Value("""float32""" ) )} ) # optional, save as float32 instead of float64 to save space _lowerCAmelCase =dataset.map( partial(__UpperCamelCase , ctx_encoder=__UpperCamelCase , ctx_tokenizer=__UpperCamelCase ) , batched=__UpperCamelCase , batch_size=processing_args.batch_size , features=__UpperCamelCase , ) # And finally save your dataset _lowerCAmelCase =os.path.join(rag_example_args.output_dir , """my_knowledge_dataset""" ) dataset.save_to_disk(__UpperCamelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("""Step 2 - Index the dataset""" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _lowerCAmelCase =faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("""embeddings""" , custom_index=__UpperCamelCase ) # And save the index _lowerCAmelCase =os.path.join(rag_example_args.output_dir , """my_knowledge_dataset_hnsw_index.faiss""" ) dataset.get_index("""embeddings""" ).save(__UpperCamelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCamelCase__ : '''simple docstring''' lowerCamelCase = field( default=str(Path(__magic_name__ ).parent / '''test_run''' / '''dummy-kb''' / '''my_knowledge_dataset.csv''' ) , metadata={'''help''': '''Path to a tab-separated csv file with columns \'title\' and \'text\''''} , ) lowerCamelCase = field( default=__magic_name__ , metadata={'''help''': '''Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'''} , ) lowerCamelCase = field( default='''facebook/rag-sequence-nq''' , metadata={'''help''': '''The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''''} , ) lowerCamelCase = field( default='''facebook/dpr-ctx_encoder-multiset-base''' , metadata={ '''help''': ( '''The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or''' ''' \'facebook/dpr-ctx_encoder-multiset-base\'''' ) } , ) lowerCamelCase = field( default=str(Path(__magic_name__ ).parent / '''test_run''' / '''dummy-kb''' ) , metadata={'''help''': '''Path to a directory where the dataset passages and the index will be saved'''} , ) @dataclass class lowerCamelCase__ : '''simple docstring''' lowerCamelCase = field( default=__magic_name__ , metadata={ '''help''': '''The number of processes to use to split the documents into passages. Default is single process.''' } , ) lowerCamelCase = field( default=16 , metadata={ '''help''': '''The batch size to use when computing the passages embeddings using the DPR context encoder.''' } , ) @dataclass class lowerCamelCase__ : '''simple docstring''' lowerCamelCase = field( default=768 , metadata={'''help''': '''The dimension of the embeddings to pass to the HNSW Faiss index.'''} , ) lowerCamelCase = field( default=128 , metadata={ '''help''': ( '''The number of bi-directional links created for every new element during the HNSW index construction.''' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) __A = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) __A , __A , __A = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: __A = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
356
"""simple docstring""" def _lowerCamelCase(__UpperCamelCase ) -> Optional[Any]: _lowerCAmelCase =0 _lowerCAmelCase =len(__UpperCamelCase ) for i in range(n - 1 ): for j in range(i + 1 , __UpperCamelCase ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def _lowerCamelCase(__UpperCamelCase ) -> List[Any]: if len(__UpperCamelCase ) <= 1: return arr, 0 _lowerCAmelCase =len(__UpperCamelCase ) // 2 _lowerCAmelCase =arr[0:mid] _lowerCAmelCase =arr[mid:] _lowerCAmelCase , _lowerCAmelCase =count_inversions_recursive(__UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =count_inversions_recursive(__UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =_count_cross_inversions(__UpperCamelCase , __UpperCamelCase ) _lowerCAmelCase =inversion_p + inversions_q + cross_inversions return c, num_inversions def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase ) -> Any: _lowerCAmelCase =[] _lowerCAmelCase =_lowerCAmelCase =_lowerCAmelCase =0 while i < len(__UpperCamelCase ) and j < len(__UpperCamelCase ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__UpperCamelCase ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__UpperCamelCase ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def _lowerCamelCase() -> str: _lowerCAmelCase =[10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) _lowerCAmelCase =count_inversions_bf(__UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =count_inversions_recursive(__UpperCamelCase ) assert num_inversions_bf == num_inversions_recursive == 8 print("""number of inversions = """ , __UpperCamelCase ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() _lowerCAmelCase =count_inversions_bf(__UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =count_inversions_recursive(__UpperCamelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print("""number of inversions = """ , __UpperCamelCase ) # an empty list should also have zero inversions _lowerCAmelCase =[] _lowerCAmelCase =count_inversions_bf(__UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =count_inversions_recursive(__UpperCamelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print("""number of inversions = """ , __UpperCamelCase ) if __name__ == "__main__": main()
341
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class A ( metaclass=SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A = ["torch", "scipy"] def __init__(self , *_UpperCAmelCase , **_UpperCAmelCase ) -> List[str]: requires_backends(self , ["torch", "scipy"] ) @classmethod def a_ (cls , *_UpperCAmelCase , **_UpperCAmelCase ) -> Tuple: requires_backends(cls , ["torch", "scipy"] ) @classmethod def a_ (cls , *_UpperCAmelCase , **_UpperCAmelCase ) -> str: requires_backends(cls , ["torch", "scipy"] )
298
'''simple docstring''' import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py _lowerCAmelCase = '''src/transformers''' _lowerCAmelCase = '''docs/source/en/tasks''' def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ ): with open(snake_case__ , "r" , encoding="utf-8" , newline="\n" ) as f: __UpperCamelCase : str = f.readlines() # Find the start prompt. __UpperCamelCase : Dict = 0 while not lines[start_index].startswith(snake_case__ ): start_index += 1 start_index += 1 __UpperCamelCase : Dict = start_index while not lines[end_index].startswith(snake_case__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. _lowerCAmelCase = direct_transformers_import(TRANSFORMERS_PATH) _lowerCAmelCase = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). _lowerCAmelCase = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def __lowerCAmelCase ( snake_case__ ): __UpperCamelCase : Optional[Any] = TASK_GUIDE_TO_MODELS[task_guide] __UpperCamelCase : str = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(snake_case__ , set() ) __UpperCamelCase : Union[str, Any] = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"[{name}](../model_doc/{code})" for code, name in model_names.items()] ) + "\n" def __lowerCAmelCase ( snake_case__ , snake_case__=False ): __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : Union[str, Any] = _find_text_in_file( filename=os.path.join(snake_case__ , snake_case__ ) , start_prompt="<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->" , end_prompt="<!--End of the generated tip-->" , ) __UpperCamelCase : List[str] = get_model_list_for_task(snake_case__ ) if current_list != new_list: if overwrite: with open(os.path.join(snake_case__ , snake_case__ ) , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`" " to fix this." ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowerCAmelCase = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
298
1
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCamelCase : """simple docstring""" def __init__( self : List[str] , __magic_name__ : Optional[int] , __magic_name__ : int=13 , __magic_name__ : str=30 , __magic_name__ : Tuple=2 , __magic_name__ : Tuple=3 , __magic_name__ : Optional[Any]=True , __magic_name__ : List[str]=True , __magic_name__ : Dict=32 , __magic_name__ : List[Any]=5 , __magic_name__ : Optional[int]=4 , __magic_name__ : List[Any]=37 , __magic_name__ : Any="gelu" , __magic_name__ : List[str]=0.1 , __magic_name__ : Union[str, Any]=0.1 , __magic_name__ : Optional[Any]=10 , __magic_name__ : List[Any]=0.02 , __magic_name__ : Optional[int]=3 , __magic_name__ : List[str]=None , __magic_name__ : Any=2 , ) -> List[str]: SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = image_size SCREAMING_SNAKE_CASE_ = patch_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = type_sequence_label_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = scope SCREAMING_SNAKE_CASE_ = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) SCREAMING_SNAKE_CASE_ = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE_ = num_patches + 2 def __A ( self : Tuple ) -> Dict: SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = None if self.use_labels: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, pixel_values, labels def __A ( self : Dict ) -> Union[str, Any]: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __A ( self : int , __magic_name__ : List[str] , __magic_name__ : Optional[int] , __magic_name__ : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE_ = DeiTModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self : Tuple , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any] , __magic_name__ : Dict ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = DeiTForMaskedImageModeling(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = DeiTForMaskedImageModeling(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __A ( self : Optional[int] , __magic_name__ : Tuple , __magic_name__ : Union[str, Any] , __magic_name__ : List[str] ) -> int: SCREAMING_SNAKE_CASE_ = self.type_sequence_label_size SCREAMING_SNAKE_CASE_ = DeiTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = DeiTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __A ( self : List[str] ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() ( SCREAMING_SNAKE_CASE_ ) = config_and_inputs SCREAMING_SNAKE_CASE_ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase (_a , _a , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase__ = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def __A ( self : Any ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = DeiTModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def __A ( self : str ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def __A ( self : str ) -> List[Any]: pass def __A ( self : Any ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , nn.Linear ) ) def __A ( self : Dict ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def __A ( self : List[str] ) -> Dict: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def __A ( self : str ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def __A ( self : List[str] ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def __A ( self : Union[str, Any] , __magic_name__ : Optional[int] , __magic_name__ : Optional[int] , __magic_name__ : Union[str, Any]=False ) -> Dict: SCREAMING_SNAKE_CASE_ = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __A ( self : List[Any] ) -> Optional[int]: if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(__lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue SCREAMING_SNAKE_CASE_ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() SCREAMING_SNAKE_CASE_ = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = model(**__lowerCamelCase ).loss loss.backward() def __A ( self : Any ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = True for model_class in self.all_model_classes: if model_class in get_values(__lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue SCREAMING_SNAKE_CASE_ = model_class(__lowerCamelCase ) model.gradient_checkpointing_enable() model.to(__lowerCamelCase ) model.train() SCREAMING_SNAKE_CASE_ = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = model(**__lowerCamelCase ).loss loss.backward() def __A ( self : Any ) -> Any: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(__lowerCamelCase ), *get_values(__lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type["title"]}''' ): SCREAMING_SNAKE_CASE_ = problem_type["""title"""] SCREAMING_SNAKE_CASE_ = problem_type["""num_labels"""] SCREAMING_SNAKE_CASE_ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() SCREAMING_SNAKE_CASE_ = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if problem_type["num_labels"] > 1: SCREAMING_SNAKE_CASE_ = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) SCREAMING_SNAKE_CASE_ = inputs["""labels"""].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=__lowerCamelCase ) as warning_list: SCREAMING_SNAKE_CASE_ = model(**__lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def __A ( self : Dict ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ = DeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def a__ ( ): SCREAMING_SNAKE_CASE_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase (unittest.TestCase ): """simple docstring""" @cached_property def __A ( self : int ) -> int: return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def __A ( self : Optional[Any] ) -> Any: SCREAMING_SNAKE_CASE_ = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( __lowerCamelCase ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=__lowerCamelCase , return_tensors="pt" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**__lowerCamelCase ) # verify the logits SCREAMING_SNAKE_CASE_ = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE_ = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def __A ( self : Tuple ) -> int: SCREAMING_SNAKE_CASE_ = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=__lowerCamelCase , return_tensors="pt" ) SCREAMING_SNAKE_CASE_ = inputs.pixel_values.to(__lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(__lowerCamelCase )
368
from ....utils import logging A : List[str] = logging.get_logger(__name__) class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self : List[str] , __magic_name__ : Optional[Any] , __magic_name__ : Any=None , __magic_name__ : List[str]=2_048 ) -> List[Any]: SCREAMING_SNAKE_CASE_ = config.__dict__ SCREAMING_SNAKE_CASE_ = modal_hidden_size if num_labels: SCREAMING_SNAKE_CASE_ = num_labels
305
0
from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
204
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def a__ ( self : int ) -> int: """simple docstring""" lowerCamelCase_ = 1 lowerCamelCase_ = 3 lowerCamelCase_ = (32, 32) lowerCamelCase_ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(A_ ) return image @property def a__ ( self : Any ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) lowerCamelCase_ = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=A_ , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def a__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" torch.manual_seed(0 ) lowerCamelCase_ = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) return model @property def a__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) lowerCamelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) return CLIPTextModel(A_ ) def a__ ( self : Optional[int] ) -> List[str]: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.dummy_cond_unet_upscale lowerCamelCase_ = DDPMScheduler() lowerCamelCase_ = DDIMScheduler(prediction_type='v_prediction' ) lowerCamelCase_ = self.dummy_vae lowerCamelCase_ = self.dummy_text_encoder lowerCamelCase_ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCamelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase_ = Image.fromarray(np.uinta(A_ ) ).convert('RGB' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk lowerCamelCase_ = StableDiffusionUpscalePipeline( unet=A_ , low_res_scheduler=A_ , scheduler=A_ , vae=A_ , text_encoder=A_ , tokenizer=A_ , max_noise_level=350 , ) lowerCamelCase_ = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) lowerCamelCase_ = 'A painting of a squirrel eating a burger' lowerCamelCase_ = torch.Generator(device=A_ ).manual_seed(0 ) lowerCamelCase_ = sd_pipe( [prompt] , image=A_ , generator=A_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) lowerCamelCase_ = output.images lowerCamelCase_ = torch.Generator(device=A_ ).manual_seed(0 ) lowerCamelCase_ = sd_pipe( [prompt] , image=A_ , generator=A_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , return_dict=A_ , )[0] lowerCamelCase_ = image[0, -3:, -3:, -1] lowerCamelCase_ = image_from_tuple[0, -3:, -3:, -1] lowerCamelCase_ = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) lowerCamelCase_ = np.array([0.3113, 0.3910, 0.4272, 0.4859, 0.5061, 0.4652, 0.5362, 0.5715, 0.5661] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.dummy_cond_unet_upscale lowerCamelCase_ = DDPMScheduler() lowerCamelCase_ = DDIMScheduler(prediction_type='v_prediction' ) lowerCamelCase_ = self.dummy_vae lowerCamelCase_ = self.dummy_text_encoder lowerCamelCase_ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCamelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase_ = Image.fromarray(np.uinta(A_ ) ).convert('RGB' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk lowerCamelCase_ = StableDiffusionUpscalePipeline( unet=A_ , low_res_scheduler=A_ , scheduler=A_ , vae=A_ , text_encoder=A_ , tokenizer=A_ , max_noise_level=350 , ) lowerCamelCase_ = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) lowerCamelCase_ = 'A painting of a squirrel eating a burger' lowerCamelCase_ = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) lowerCamelCase_ = output.images assert image.shape[0] == 2 lowerCamelCase_ = torch.Generator(device=A_ ).manual_seed(0 ) lowerCamelCase_ = sd_pipe( [prompt] , image=A_ , generator=A_ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) lowerCamelCase_ = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = self.dummy_cond_unet_upscale lowerCamelCase_ = DDPMScheduler() lowerCamelCase_ = DDIMScheduler(prediction_type='v_prediction' ) lowerCamelCase_ = self.dummy_vae lowerCamelCase_ = self.dummy_text_encoder lowerCamelCase_ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCamelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase_ = Image.fromarray(np.uinta(A_ ) ).convert('RGB' ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 lowerCamelCase_ = unet.half() lowerCamelCase_ = text_encoder.half() # make sure here that pndm scheduler skips prk lowerCamelCase_ = StableDiffusionUpscalePipeline( unet=A_ , low_res_scheduler=A_ , scheduler=A_ , vae=A_ , text_encoder=A_ , tokenizer=A_ , max_noise_level=350 , ) lowerCamelCase_ = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) lowerCamelCase_ = 'A painting of a squirrel eating a burger' lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = sd_pipe( [prompt] , image=A_ , generator=A_ , num_inference_steps=2 , output_type='np' , ).images lowerCamelCase_ = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Tuple ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) lowerCamelCase_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale' '/upsampled_cat.npy' ) lowerCamelCase_ = 'stabilityai/stable-diffusion-x4-upscaler' lowerCamelCase_ = StableDiffusionUpscalePipeline.from_pretrained(A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() lowerCamelCase_ = 'a cat sitting on a park bench' lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=A_ , image=A_ , generator=A_ , output_type='np' , ) lowerCamelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-3 def a__ ( self : Dict ) -> List[str]: """simple docstring""" lowerCamelCase_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) lowerCamelCase_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale' '/upsampled_cat_fp16.npy' ) lowerCamelCase_ = 'stabilityai/stable-diffusion-x4-upscaler' lowerCamelCase_ = StableDiffusionUpscalePipeline.from_pretrained( A_ , torch_dtype=torch.floataa , ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() lowerCamelCase_ = 'a cat sitting on a park bench' lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=A_ , image=A_ , generator=A_ , output_type='np' , ) lowerCamelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def a__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) lowerCamelCase_ = 'stabilityai/stable-diffusion-x4-upscaler' lowerCamelCase_ = StableDiffusionUpscalePipeline.from_pretrained( A_ , torch_dtype=torch.floataa , ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCamelCase_ = 'a cat sitting on a park bench' lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = pipe( prompt=A_ , image=A_ , generator=A_ , num_inference_steps=5 , output_type='np' , ) lowerCamelCase_ = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
204
1
def A ( _UpperCAmelCase : str , _UpperCAmelCase : str , _UpperCAmelCase : Any , _UpperCAmelCase : Tuple ) -> Tuple: '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: _UpperCAmelCase = mf_knapsack(i - 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: _UpperCAmelCase = max( mf_knapsack(i - 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , mf_knapsack(i - 1 , _UpperCAmelCase , _UpperCAmelCase , j - wt[i - 1] ) + val[i - 1] , ) _UpperCAmelCase = val return f[i][j] def A ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : int , _UpperCAmelCase : Any ) -> Tuple: '''simple docstring''' _UpperCAmelCase = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: _UpperCAmelCase = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: _UpperCAmelCase = dp[i - 1][w_] return dp[n][w_], dp def A ( _UpperCAmelCase : int , _UpperCAmelCase : list , _UpperCAmelCase : list ) -> Union[str, Any]: '''simple docstring''' if not (isinstance(_UpperCAmelCase , (list, tuple) ) and isinstance(_UpperCAmelCase , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) _UpperCAmelCase = len(_UpperCAmelCase ) if num_items != len(_UpperCAmelCase ): _UpperCAmelCase = ( 'The number of weights must be the same as the number of values.\n' F"But got {num_items} weights and {len(_UpperCAmelCase )} values" ) raise ValueError(_UpperCAmelCase ) for i in range(_UpperCAmelCase ): if not isinstance(wt[i] , _UpperCAmelCase ): _UpperCAmelCase = ( 'All weights must be integers but got weight of ' F"type {type(wt[i] )} at index {i}" ) raise TypeError(_UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase = knapsack(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) _UpperCAmelCase = set() _construct_solution(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return optimal_val, example_optional_set def A ( _UpperCAmelCase : list , _UpperCAmelCase : list , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : set ) -> int: '''simple docstring''' # for the current item i at a maximum weight j to be part of an optimal subset, # the optimal value at (i, j) must be greater than the optimal value at (i-1, j). # where i - 1 means considering only the previous items at the given maximum weight if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(_UpperCAmelCase , _UpperCAmelCase , i - 1 , _UpperCAmelCase , _UpperCAmelCase ) else: optimal_set.add(_UpperCAmelCase ) _construct_solution(_UpperCAmelCase , _UpperCAmelCase , i - 1 , j - wt[i - 1] , _UpperCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ = [3, 2, 4, 4] UpperCAmelCase__ = [4, 3, 2, 3] UpperCAmelCase__ = 4 UpperCAmelCase__ = 6 UpperCAmelCase__ = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] UpperCAmelCase__ , UpperCAmelCase__ = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 UpperCAmelCase__ , UpperCAmelCase__ = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("optimal_value = ", optimal_solution) print("An optimal subset corresponding to the optimal value", optimal_subset)
362
UpperCAmelCase__ = { "A": ["B", "C", "E"], "B": ["A", "D", "E"], "C": ["A", "F", "G"], "D": ["B"], "E": ["A", "B", "D"], "F": ["C"], "G": ["C"], } def A ( _UpperCAmelCase : dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[Any] ) -> list[str]: '''simple docstring''' _UpperCAmelCase = set() # keep track of all the paths to be checked _UpperCAmelCase = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue _UpperCAmelCase = queue.pop(0 ) # get the last node from the path _UpperCAmelCase = path[-1] if node not in explored: _UpperCAmelCase = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: _UpperCAmelCase = list(_UpperCAmelCase ) new_path.append(_UpperCAmelCase ) queue.append(_UpperCAmelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_UpperCAmelCase ) # in case there's no path between the 2 nodes return [] def A ( _UpperCAmelCase : dict , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] ) -> int: '''simple docstring''' if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 _UpperCAmelCase = [start] _UpperCAmelCase = set(_UpperCAmelCase ) # Keep tab on distances from `start` node. _UpperCAmelCase = {start: 0, target: -1} while queue: _UpperCAmelCase = queue.pop(0 ) if node == target: _UpperCAmelCase = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_UpperCAmelCase ) queue.append(_UpperCAmelCase ) _UpperCAmelCase = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, "G", "D")) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, "G", "D")) # returns 4
290
0
def a ( _UpperCAmelCase : int ): '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
226
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __A =logging.get_logger(__name__) def a ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict ): '''simple docstring''' __UpperCAmelCase : List[str] = b.T __UpperCAmelCase : Any = np.sum(np.square(_UpperCAmelCase ) , axis=1 ) __UpperCAmelCase : int = np.sum(np.square(_UpperCAmelCase ) , axis=0 ) __UpperCAmelCase : Optional[int] = np.matmul(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase : List[str] = aa[:, None] - 2 * ab + ba[None, :] return d def a ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : str = x.reshape(-1 , 3 ) __UpperCAmelCase : Optional[int] = squared_euclidean_distance(_UpperCAmelCase , _UpperCAmelCase ) return np.argmin(_UpperCAmelCase , axis=1 ) class UpperCAmelCase__ ( __UpperCamelCase ): '''simple docstring''' UpperCamelCase = ["""pixel_values"""] def __init__( self : str , a_ : Optional[Union[List[List[int]], np.ndarray]] = None , a_ : bool = True , a_ : Dict[str, int] = None , a_ : PILImageResampling = PILImageResampling.BILINEAR , a_ : bool = True , a_ : bool = True , **a_ : List[str] , ): '''simple docstring''' super().__init__(**a_ ) __UpperCAmelCase : Optional[int] = size if size is not None else {'''height''': 2_56, '''width''': 2_56} __UpperCAmelCase : List[str] = get_size_dict(a_ ) __UpperCAmelCase : str = np.array(a_ ) if clusters is not None else None __UpperCAmelCase : Dict = do_resize __UpperCAmelCase : Tuple = size __UpperCAmelCase : Union[str, Any] = resample __UpperCAmelCase : Tuple = do_normalize __UpperCAmelCase : Optional[int] = do_color_quantize def snake_case__ ( self : Optional[Any] , a_ : np.ndarray , a_ : Dict[str, int] , a_ : PILImageResampling = PILImageResampling.BILINEAR , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Dict , ): '''simple docstring''' __UpperCAmelCase : Tuple = get_size_dict(a_ ) if "height" not in size or "width" not in size: raise ValueError(F'Size dictionary must contain both height and width keys. Got {size.keys()}' ) return resize( a_ , size=(size['''height'''], size['''width''']) , resample=a_ , data_format=a_ , **a_ ) def snake_case__ ( self : Tuple , a_ : np.ndarray , a_ : Optional[Union[str, ChannelDimension]] = None , ): '''simple docstring''' __UpperCAmelCase : Dict = rescale(image=a_ , scale=1 / 1_2_7.5 , data_format=a_ ) __UpperCAmelCase : Union[str, Any] = image - 1 return image def snake_case__ ( self : int , a_ : ImageInput , a_ : bool = None , a_ : Dict[str, int] = None , a_ : PILImageResampling = None , a_ : bool = None , a_ : Optional[bool] = None , a_ : Optional[Union[List[List[int]], np.ndarray]] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[Union[str, ChannelDimension]] = ChannelDimension.FIRST , **a_ : Any , ): '''simple docstring''' __UpperCAmelCase : Any = do_resize if do_resize is not None else self.do_resize __UpperCAmelCase : List[str] = size if size is not None else self.size __UpperCAmelCase : Any = get_size_dict(a_ ) __UpperCAmelCase : Optional[int] = resample if resample is not None else self.resample __UpperCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize __UpperCAmelCase : int = do_color_quantize if do_color_quantize is not None else self.do_color_quantize __UpperCAmelCase : Optional[int] = clusters if clusters is not None else self.clusters __UpperCAmelCase : Any = np.array(a_ ) __UpperCAmelCase : Optional[int] = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_color_quantize and clusters is None: raise ValueError('''Clusters must be specified if do_color_quantize is True.''' ) # All transformations expect numpy arrays. __UpperCAmelCase : List[Any] = [to_numpy_array(a_ ) for image in images] if do_resize: __UpperCAmelCase : List[str] = [self.resize(image=a_ , size=a_ , resample=a_ ) for image in images] if do_normalize: __UpperCAmelCase : Dict = [self.normalize(image=a_ ) for image in images] if do_color_quantize: __UpperCAmelCase : int = [to_channel_dimension_format(a_ , ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) __UpperCAmelCase : List[str] = np.array(a_ ) __UpperCAmelCase : Dict = color_quantize(a_ , a_ ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) __UpperCAmelCase : Any = images.shape[0] __UpperCAmelCase : Any = images.reshape(a_ , -1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. __UpperCAmelCase : List[Any] = list(a_ ) else: __UpperCAmelCase : int = [to_channel_dimension_format(a_ , a_ ) for image in images] __UpperCAmelCase : int = {'''input_ids''': images} return BatchFeature(data=a_ , tensor_type=a_ )
226
1
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class a__ ( __snake_case , __snake_case ): A__ : Tuple = 1 @register_to_config def __init__( self , UpperCAmelCase = 1_0_0_0 , UpperCAmelCase = None ) -> List[str]: # set `betas`, `alphas`, `timesteps` self.set_timesteps(UpperCAmelCase ) # standard deviation of the initial noise distribution __a = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __a = 4 # running values __a = [] def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase = None ) -> List[Any]: __a = num_inference_steps __a = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __a = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __a = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __a = torch.sin(steps * math.pi / 2 ) ** 2 __a = (1.0 - self.betas**2) ** 0.5 __a = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __a = timesteps.to(UpperCAmelCase ) __a = [] def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) __a = (self.timesteps == timestep).nonzero().item() __a = timestep_index + 1 __a = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCAmelCase ) if len(self.ets ) == 1: __a = self.ets[-1] elif len(self.ets ) == 2: __a = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __a = (2_3 * self.ets[-1] - 1_6 * self.ets[-2] + 5 * self.ets[-3]) / 1_2 else: __a = (1 / 2_4) * (5_5 * self.ets[-1] - 5_9 * self.ets[-2] + 3_7 * self.ets[-3] - 9 * self.ets[-4]) __a = self._get_prev_sample(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) -> torch.FloatTensor: return sample def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: __a = self.alphas[timestep_index] __a = self.betas[timestep_index] __a = self.alphas[prev_timestep_index] __a = self.betas[prev_timestep_index] __a = (sample - sigma * ets) / max(UpperCAmelCase , 1e-8 ) __a = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ) -> Any: return self.config.num_train_timesteps
197
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class a__ ( unittest.TestCase ): def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase ) -> Optional[Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): __a = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: __a = 'sshleifer/tiny-gpt2' __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCAmelCase , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = 'sgugger/tiny-distilbert-classification' __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , only_pretrain_model=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: __a = 'sshleifer/tiny-gpt2' __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = 'sshleifer/tiny-gpt2' __a = AutoConfig.from_pretrained(UpperCAmelCase ) __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCAmelCase , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase , [config] ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: __a = 'sshleifer/tiny-gpt2' __a = AutoConfig.from_pretrained(UpperCAmelCase ) __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase , [config] ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = 'sshleifer/tiny-gpt2' __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: __a = 'sshleifer/tiny-gpt2' __a = AutoConfig.from_pretrained(UpperCAmelCase ) __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase , [config] ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: __a = 'patrickvonplaten/t5-tiny-random' __a = AutoConfig.from_pretrained(UpperCAmelCase ) __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase , configs=[config] ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('GPU' ) ) == 0 , 'Cannot do xla on CPU.' ) def __SCREAMING_SNAKE_CASE ( self ) -> str: __a = 'sshleifer/tiny-gpt2' __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , use_xla=UpperCAmelCase , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __SCREAMING_SNAKE_CASE ( self ) -> Dict: __a = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCAmelCase , save_to_csv=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCAmelCase , 'inf_time.csv' ) , inference_memory_csv_file=os.path.join(UpperCAmelCase , 'inf_mem.csv' ) , env_info_csv_file=os.path.join(UpperCAmelCase , 'env.csv' ) , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , 'env.csv' ) ).exists() ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: __a = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(UpperCAmelCase ): self.assertTrue(hasattr(UpperCAmelCase , 'sequential' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'cumulative' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'current' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: __a = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCAmelCase , 'log.txt' ) , log_print=UpperCAmelCase , trace_memory_line_by_line=UpperCAmelCase , eager_mode=UpperCAmelCase , multi_process=UpperCAmelCase , ) __a = TensorFlowBenchmark(UpperCAmelCase ) __a = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase , 'log.txt' ) ).exists() )
197
1
"""simple docstring""" from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _UpperCAmelCase : '''simple docstring''' pass
102
"""simple docstring""" import numpy as np def lowercase ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) ->Dict: """simple docstring""" __snake_case : Union[str, Any] = int(np.ceil((x_end - xa) / h ) ) __snake_case : Dict = np.zeros((n + 1,) ) __snake_case : List[Any] = ya __snake_case : int = xa for k in range(_snake_case ): __snake_case : Any = f(_snake_case , y[k] ) __snake_case : List[Any] = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) __snake_case : int = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) __snake_case : Optional[int] = f(x + h , y[k] + h * ka ) __snake_case : Optional[int] = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
102
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __snake_case : List[str] = logging.get_logger(__name__) if is_vision_available(): import PIL class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : List[str] = ['''pixel_values'''] def __init__( self , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = PILImageResampling.BICUBIC , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = True , _UpperCamelCase = 1 / 2_55 , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = True , **_UpperCamelCase , ): """simple docstring""" super().__init__(**_UpperCamelCase ) lowerCAmelCase__ = size if size is not None else {'shortest_edge': 2_24} lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase , param_name='crop_size' ) lowerCAmelCase__ = do_resize lowerCAmelCase__ = size lowerCAmelCase__ = resample lowerCAmelCase__ = do_center_crop lowerCAmelCase__ = crop_size lowerCAmelCase__ = do_rescale lowerCAmelCase__ = rescale_factor lowerCAmelCase__ = do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowerCAmelCase__ = image_std if image_std is not None else OPENAI_CLIP_STD lowerCAmelCase__ = do_convert_rgb def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = PILImageResampling.BICUBIC , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) lowerCAmelCase__ = get_resize_output_image_size(_UpperCamelCase , size=size['shortest_edge'] , default_to_square=_UpperCamelCase ) return resize(_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(_UpperCamelCase , size=(size['height'], size['width']) , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = ChannelDimension.FIRST , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = do_resize if do_resize is not None else self.do_resize lowerCAmelCase__ = size if size is not None else self.size lowerCAmelCase__ = get_size_dict(_UpperCamelCase , param_name='size' , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = resample if resample is not None else self.resample lowerCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCAmelCase__ = crop_size if crop_size is not None else self.crop_size lowerCAmelCase__ = get_size_dict(_UpperCamelCase , param_name='crop_size' , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else self.image_mean lowerCAmelCase__ = image_std if image_std is not None else self.image_std lowerCAmelCase__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCAmelCase__ = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCAmelCase__ = [convert_to_rgb(_UpperCamelCase ) for image in images] # All transformations expect numpy arrays. lowerCAmelCase__ = [to_numpy_array(_UpperCamelCase ) for image in images] if do_resize: lowerCAmelCase__ = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] if do_center_crop: lowerCAmelCase__ = [self.center_crop(image=_UpperCamelCase , size=_UpperCamelCase ) for image in images] if do_rescale: lowerCAmelCase__ = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase ) for image in images] if do_normalize: lowerCAmelCase__ = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase ) for image in images] lowerCAmelCase__ = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] lowerCAmelCase__ = {'pixel_values': images} return BatchFeature(data=_UpperCamelCase , tensor_type=_UpperCamelCase )
122
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : Dict = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : str = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys __snake_case : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
122
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = "▁" UpperCAmelCase__ = {"vocab_file": "sentencepiece.bpe.model"} UpperCAmelCase__ = { "vocab_file": { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model" ), } } UpperCAmelCase__ = { "xlm-roberta-base": 512, "xlm-roberta-large": 512, "xlm-roberta-large-finetuned-conll02-dutch": 512, "xlm-roberta-large-finetuned-conll02-spanish": 512, "xlm-roberta-large-finetuned-conll03-english": 512, "xlm-roberta-large-finetuned-conll03-german": 512, } class lowercase_ ( lowercase ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any]="<s>" , __UpperCAmelCase : Optional[int]="</s>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<s>" , __UpperCAmelCase : Union[str, Any]="<unk>" , __UpperCAmelCase : Optional[Any]="<pad>" , __UpperCAmelCase : str="<mask>" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Any , ) ->None: """simple docstring""" a = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token a = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCAmelCase ) ) a = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token a = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab a = 1 a = len(self.sp_model ) + self.fairseq_offset a = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[str] ) ->Optional[int]: """simple docstring""" a = self.__dict__.copy() a = None a = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , __UpperCAmelCase : Tuple ) ->Optional[Any]: """simple docstring""" a = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): a = {} a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) ->List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a = [self.cls_token_id] a = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowerCAmelCase ( self : str , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) ->List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] + ([0] * len(__UpperCAmelCase )) + [1] def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) ->List[int]: """simple docstring""" a = [self.sep_token_id] a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowerCAmelCase ( self : Tuple ) ->List[Any]: """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def __lowerCAmelCase ( self : Dict ) ->Tuple: """simple docstring""" a = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : str ) ->List[str]: """simple docstring""" return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : Optional[int] ) ->Optional[int]: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] a = self.sp_model.PieceToId(__UpperCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __lowerCAmelCase ( self : Dict , __UpperCAmelCase : str ) ->List[str]: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __lowerCAmelCase ( self : Tuple , __UpperCAmelCase : Optional[Any] ) ->Union[str, Any]: """simple docstring""" a = ''''''.join(__UpperCAmelCase ).replace(__UpperCAmelCase , ''' ''' ).strip() return out_string def __lowerCAmelCase ( self : Any , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) ->Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return a = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: a = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : int = logging.get_logger(__name__) lowerCamelCase : List[Any] = { 'google/realm-cc-news-pretrained-embedder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-encoder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-scorer': ( 'https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-openqa': ( 'https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json' ), 'google/realm-orqa-nq-openqa': 'https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json', 'google/realm-orqa-nq-reader': 'https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json', 'google/realm-orqa-wq-openqa': 'https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json', 'google/realm-orqa-wq-reader': 'https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json', # See all REALM models at https://huggingface.co/models?filter=realm } class __lowerCAmelCase (lowercase_ ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = """realm""" def __init__(self : str , UpperCamelCase : List[Any]=30522 , UpperCamelCase : List[Any]=768 , UpperCamelCase : int=128 , UpperCamelCase : Any=12 , UpperCamelCase : Tuple=12 , UpperCamelCase : List[Any]=8 , UpperCamelCase : Union[str, Any]=3072 , UpperCamelCase : List[str]="gelu_new" , UpperCamelCase : Any=0.1 , UpperCamelCase : List[str]=0.1 , UpperCamelCase : Dict=512 , UpperCamelCase : Dict=2 , UpperCamelCase : List[Any]=0.02 , UpperCamelCase : List[Any]=1E-12 , UpperCamelCase : Dict=256 , UpperCamelCase : Union[str, Any]=10 , UpperCamelCase : Optional[int]=1E-3 , UpperCamelCase : Tuple=5 , UpperCamelCase : Optional[int]=320 , UpperCamelCase : List[str]=13353718 , UpperCamelCase : Optional[Any]=5000 , UpperCamelCase : str=1 , UpperCamelCase : Union[str, Any]=0 , UpperCamelCase : List[Any]=2 , **UpperCamelCase : int , ): '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase , bos_token_id=UpperCamelCase , eos_token_id=UpperCamelCase , **UpperCamelCase ) # Common config lowercase__ = vocab_size lowercase__ = max_position_embeddings lowercase__ = hidden_size lowercase__ = retriever_proj_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_candidates lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = initializer_range lowercase__ = type_vocab_size lowercase__ = layer_norm_eps # Reader config lowercase__ = span_hidden_size lowercase__ = max_span_width lowercase__ = reader_layer_norm_eps lowercase__ = reader_beam_size lowercase__ = reader_seq_len # Retrieval config lowercase__ = num_block_records lowercase__ = searcher_beam_size
2
0
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A__ : def __init__( self : Optional[int] , a : Dict , a : Tuple=13 , a : Union[str, Any]=30 , a : int=2 , a : Dict=3 , a : Optional[int]=True , a : Dict=True , a : Union[str, Any]=32 , a : List[Any]=5 , a : str=4 , a : Optional[int]=37 , a : Tuple="gelu" , a : Optional[int]=0.1 , a : int=0.1 , a : List[str]=10 , a : str=0.0_2 , a : Union[str, Any]=None , a : Optional[int]=2 , ): '''simple docstring''' lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : Tuple = image_size lowerCAmelCase__ : List[Any] = patch_size lowerCAmelCase__ : List[str] = num_channels lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : int = use_labels lowerCAmelCase__ : Optional[int] = hidden_size lowerCAmelCase__ : Tuple = num_hidden_layers lowerCAmelCase__ : Optional[int] = num_attention_heads lowerCAmelCase__ : str = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = type_sequence_label_size lowerCAmelCase__ : Dict = initializer_range lowerCAmelCase__ : Any = scope lowerCAmelCase__ : str = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : Any = (image_size // patch_size) ** 2 lowerCAmelCase__ : Optional[int] = num_patches + 1 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Dict = None if self.use_labels: lowerCAmelCase__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : List[Any] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : Dict ): '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=a , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[Any] , a : int , a : List[Any] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = ViTModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : Optional[int] , a : Any , a : Dict , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = ViTForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : List[str] = ViTForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Dict = model(a ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : List[str] , a : Tuple , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.type_sequence_label_size lowerCAmelCase__ : Union[str, Any] = ViTForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : int = 1 lowerCAmelCase__ : List[str] = ViTForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : int = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Dict = config_and_inputs lowerCAmelCase__ : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase = ( {'feature-extraction': ViTModel, 'image-classification': ViTForImageClassification} if is_torch_available() else {} ) lowercase = True lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = ViTModelTester(self ) lowerCAmelCase__ : str = ConfigTester(self , config_class=a , has_text_modality=a , hidden_size=37 ) def _lowerCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' pass def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[Any] = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = model_class(a ) lowerCAmelCase__ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Optional[Any] = [*signature.parameters.keys()] lowerCAmelCase__ : Optional[int] = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[Any] = ViTModel.from_pretrained(a ) self.assertIsNotNone(a ) def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Any ): '''simple docstring''' return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = ViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ).to(a ) lowerCAmelCase__ : Optional[Any] = self.default_image_processor lowerCAmelCase__ : List[str] = prepare_img() lowerCAmelCase__ : Dict = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Dict = model(**a ) # verify the logits lowerCAmelCase__ : Tuple = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Union[str, Any] = torch.tensor([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) ) @slow def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = ViTModel.from_pretrained('facebook/dino-vits8' ).to(a ) lowerCAmelCase__ : List[Any] = ViTImageProcessor.from_pretrained('facebook/dino-vits8' , size=480 ) lowerCAmelCase__ : Dict = prepare_img() lowerCAmelCase__ : List[Any] = image_processor(images=a , return_tensors='pt' ) lowerCAmelCase__ : Optional[int] = inputs.pixel_values.to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : List[str] = model(a , interpolate_pos_encoding=a ) # verify the logits lowerCAmelCase__ : str = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , a ) lowerCAmelCase__ : int = torch.tensor( [[4.2_3_4_0, 4.3_9_0_6, -6.6_6_9_2], [4.5_4_6_3, 1.8_9_2_8, -6.7_2_5_7], [4.4_4_2_9, 0.8_4_9_6, -5.8_5_8_5]] ).to(a ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , a , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : int = ViTModel.from_pretrained('facebook/dino-vits8' , torch_dtype=torch.floataa , device_map='auto' ) lowerCAmelCase__ : str = self.default_image_processor lowerCAmelCase__ : Union[str, Any] = prepare_img() lowerCAmelCase__ : Optional[Any] = image_processor(images=a , return_tensors='pt' ) lowerCAmelCase__ : str = inputs.pixel_values.to(a ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(a )
307
import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class A__ ( __magic_name__ ): lowercase = (DPMSolverSDEScheduler,) lowercase = 10 def _lowerCamelCase ( self : Optional[int] , **a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = { 'num_train_timesteps': 1_100, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'noise_sampler_seed': 0, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.scheduler_classes[0] lowerCAmelCase__ : Tuple = self.get_scheduler_config() lowerCAmelCase__ : List[Any] = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : Dict = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : int = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : List[Any] = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : str = model(a , a ) lowerCAmelCase__ : int = scheduler.step(a , a , a ) lowerCAmelCase__ : Any = output.prev_sample lowerCAmelCase__ : List[Any] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[int] = torch.mean(torch.abs(a ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_6_7.4_7_8_2_1_0_4_4_9_2_1_8_7_5 ) < 1E-2 assert abs(result_mean.item() - 0.2_1_7_8_7_0_5_9_6_4_5_6_5_2_7_7 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_7_1.5_9_3_5_2_1_1_1_8_1_6_4_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_2_3_4_2_9_0_6_8_9_2_2_9_9_6_5_2 ) < 1E-3 else: assert abs(result_sum.item() - 1_6_2.5_2_3_8_3_4_2_2_8_5_1_5_6_2 ) < 1E-2 assert abs(result_mean.item() - 0.2_1_1_6_1_9_5_7_0_8_5_1_3_2_6 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : List[str] = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : Any = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : Any = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : str = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : str = model(a , a ) lowerCAmelCase__ : Dict = scheduler.step(a , a , a ) lowerCAmelCase__ : Tuple = output.prev_sample lowerCAmelCase__ : int = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Union[str, Any] = torch.mean(torch.abs(a ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_2_4.7_7_1_4_9_2_0_0_4_3_9_4_5_3 ) < 1E-2 assert abs(result_mean.item() - 0.1_6_2_2_6_2_8_9_0_1_4_8_1_6_2_8_4 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_2_8.1_6_6_3_3_6_0_5_9_5_7_0_3 ) < 1E-2 assert abs(result_mean.item() - 0.1_6_6_8_8_3_2_6_0_0_1_1_6_7_2_9_7 ) < 1E-3 else: assert abs(result_sum.item() - 1_1_9.8_4_8_7_5_4_8_8_2_8_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.1_5_6_0_5_3_0_6_6_2_5_3_6_6_2_1 ) < 1E-3 def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps , device=a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Any = self.dummy_sample_deter.to(a ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase__ : Dict = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : Optional[int] = model(a , a ) lowerCAmelCase__ : Tuple = scheduler.step(a , a , a ) lowerCAmelCase__ : Dict = output.prev_sample lowerCAmelCase__ : Union[str, Any] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Dict = torch.mean(torch.abs(a ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_6_7.4_6_9_5_7_3_9_7_4_6_0_9_3_8 ) < 1E-2 assert abs(result_mean.item() - 0.2_1_8_0_5_9_3_4_6_0_7_9_8_2_6_3_5 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_7_1.5_9_3_5_3_6_3_7_6_9_5_3_1_2 ) < 1E-2 assert abs(result_mean.item() - 0.2_2_3_4_2_9_0_8_3_8_2_4_1_5_7_7_1 ) < 1E-3 else: assert abs(result_sum.item() - 1_6_2.5_2_3_8_3_4_2_2_8_5_1_5_6_2 ) < 1E-2 assert abs(result_mean.item() - 0.2_1_1_6_1_9_5_7_0_8_5_1_3_2_6 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Any = scheduler_class(**a , use_karras_sigmas=a ) scheduler.set_timesteps(self.num_inference_steps , device=a ) lowerCAmelCase__ : str = self.dummy_model() lowerCAmelCase__ : Any = self.dummy_sample_deter.to(a ) * scheduler.init_noise_sigma lowerCAmelCase__ : str = sample.to(a ) for t in scheduler.timesteps: lowerCAmelCase__ : Any = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : int = model(a , a ) lowerCAmelCase__ : Union[str, Any] = scheduler.step(a , a , a ) lowerCAmelCase__ : Union[str, Any] = output.prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_7_6.6_6_9_7_4_1_3_5_7_4_2_1_8_8 ) < 1E-2 assert abs(result_mean.item() - 0.2_3_0_0_3_8_7_2_7_3_0_9_8_1_8_1_1 ) < 1E-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_7_7.6_3_6_5_3_5_6_4_4_5_3_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.2_3_0_0_3_8_7_2_7_3_0_9_8_1_8_1_1 ) < 1E-2 else: assert abs(result_sum.item() - 1_7_0.3_1_3_5_2_2_3_3_8_8_6_7_2 ) < 1E-2 assert abs(result_mean.item() - 0.2_3_0_0_3_8_7_2_7_3_0_9_8_1_8_1_1 ) < 1E-2
307
1
a_ = [ '''DownloadConfig''', '''DownloadManager''', '''DownloadMode''', '''StreamingDownloadManager''', ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
340
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase__ ( _UpperCAmelCase ): a_ =["""image_processor""", """tokenizer"""] a_ ="""LayoutLMv2ImageProcessor""" a_ =("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase )-> Tuple: '''simple docstring''' if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __UpperCAmelCase , ) lowerCAmelCase__ = kwargs.pop("feature_extractor" ) lowerCAmelCase__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def __call__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = True , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = 0 , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = None , **__UpperCAmelCase , )-> BatchEncoding: '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes " "if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError("You cannot return overflowing tokens without returning the offsets mapping." ) # first, apply the image processor lowerCAmelCase__ = self.image_processor(images=__UpperCAmelCase , return_tensors=__UpperCAmelCase ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ = features["words"] lowerCAmelCase__ = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=__UpperCAmelCase , max_length=__UpperCAmelCase , stride=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , return_overflowing_tokens=__UpperCAmelCase , return_special_tokens_mask=__UpperCAmelCase , return_offsets_mapping=__UpperCAmelCase , return_length=__UpperCAmelCase , verbose=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) # add pixel values lowerCAmelCase__ = features.pop("pixel_values" ) if return_overflowing_tokens is True: lowerCAmelCase__ = self.get_overflowing_images(__UpperCAmelCase , encoded_inputs["overflow_to_sample_mapping"] ) lowerCAmelCase__ = images return encoded_inputs def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase )-> str: '''simple docstring''' lowerCAmelCase__ = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(__UpperCAmelCase ) != len(__UpperCAmelCase ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" F" {len(__UpperCAmelCase )} and {len(__UpperCAmelCase )}" ) return images_with_overflow def UpperCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase )-> Union[str, Any]: '''simple docstring''' return self.tokenizer.batch_decode(*__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase )-> Dict: '''simple docstring''' return self.tokenizer.decode(*__UpperCAmelCase , **__UpperCAmelCase ) @property def UpperCAmelCase ( self )-> Optional[int]: '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "image"] @property def UpperCAmelCase ( self )-> Union[str, Any]: '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __UpperCAmelCase , ) return self.image_processor_class @property def UpperCAmelCase ( self )-> str: '''simple docstring''' warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __UpperCAmelCase , ) return self.image_processor
340
1
"""simple docstring""" import socket def __SCREAMING_SNAKE_CASE ( ): _lowercase : Optional[int] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowercase : Optional[int] = socket.gethostname() _lowercase : Dict = 12312 sock.connect((host, port) ) sock.send(B"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: _lowercase : List[str] = sock.recv(1024 ) if not data: break out_file.write(__UpperCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
356
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : Optional[int] = "BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Tuple = "AutoTokenizer" def __init__( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ): super().__init__(UpperCAmelCase_ ,UpperCAmelCase_ ) # add QFormer tokenizer _lowercase : Optional[int] = qformer_tokenizer def __call__( self ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = True ,UpperCAmelCase_ = False ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = 0 ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = True ,UpperCAmelCase_ = None ,**UpperCAmelCase_ ,): if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) _lowercase : List[Any] = BatchFeature() if text is not None: _lowercase : List[str] = self.tokenizer( text=UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,padding=UpperCAmelCase_ ,truncation=UpperCAmelCase_ ,max_length=UpperCAmelCase_ ,stride=UpperCAmelCase_ ,pad_to_multiple_of=UpperCAmelCase_ ,return_attention_mask=UpperCAmelCase_ ,return_overflowing_tokens=UpperCAmelCase_ ,return_special_tokens_mask=UpperCAmelCase_ ,return_offsets_mapping=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ,return_length=UpperCAmelCase_ ,verbose=UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ,**UpperCAmelCase_ ,) encoding.update(UpperCAmelCase_ ) _lowercase : Dict = self.qformer_tokenizer( text=UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,padding=UpperCAmelCase_ ,truncation=UpperCAmelCase_ ,max_length=UpperCAmelCase_ ,stride=UpperCAmelCase_ ,pad_to_multiple_of=UpperCAmelCase_ ,return_attention_mask=UpperCAmelCase_ ,return_overflowing_tokens=UpperCAmelCase_ ,return_special_tokens_mask=UpperCAmelCase_ ,return_offsets_mapping=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ,return_length=UpperCAmelCase_ ,verbose=UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ,**UpperCAmelCase_ ,) _lowercase : str = qformer_text_encoding.pop("""input_ids""" ) _lowercase : int = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: _lowercase : Optional[Any] = self.image_processor(UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ) encoding.update(UpperCAmelCase_ ) return encoding def lowerCamelCase__ ( self ,*UpperCAmelCase_ ,**UpperCAmelCase_ ): return self.tokenizer.batch_decode(*UpperCAmelCase_ ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,*UpperCAmelCase_ ,**UpperCAmelCase_ ): return self.tokenizer.decode(*UpperCAmelCase_ ,**UpperCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = self.tokenizer.model_input_names _lowercase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,**UpperCAmelCase_ ): if os.path.isfile(UpperCAmelCase_ ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(UpperCAmelCase_ ,exist_ok=UpperCAmelCase_ ) _lowercase : Any = os.path.join(UpperCAmelCase_ ,"""qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(UpperCAmelCase_ ) return super().save_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) @classmethod def lowerCamelCase__ ( cls ,UpperCAmelCase_ ,**UpperCAmelCase_ ): _lowercase : Optional[Any] = AutoTokenizer.from_pretrained(UpperCAmelCase_ ,subfolder="""qformer_tokenizer""" ) _lowercase : Any = cls._get_arguments_from_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) args.append(UpperCAmelCase_ ) return cls(*UpperCAmelCase_ )
336
0
from math import asin, atan, cos, radians, sin, sqrt, tan _snake_case = 6_3_7_8_1_3_7.0 _snake_case = 6_3_5_6_7_5_2.3_1_4_2_4_5 _snake_case = 6378137 def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_ ): _A : Any = (AXIS_A - AXIS_B) / AXIS_A _A : Optional[int] = atan((1 - flattening) * tan(radians(snake_case_ ) ) ) _A : List[str] = atan((1 - flattening) * tan(radians(snake_case_ ) ) ) _A : Optional[Any] = radians(snake_case_ ) _A : str = radians(snake_case_ ) # Equation _A : Dict = sin((phi_a - phi_a) / 2 ) _A : List[str] = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda _A : Optional[int] = sqrt(sin_sq_phi + (cos(snake_case_ ) * cos(snake_case_ ) * sin_sq_lambda) ) return 2 * RADIUS * asin(snake_case_ ) if __name__ == "__main__": import doctest doctest.testmod()
26
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class A : def __init__( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Dict=13 , lowercase_ : List[Any]=7 , lowercase_ : List[Any]=True , lowercase_ : Any=True , lowercase_ : Optional[int]=False , lowercase_ : Tuple=True , lowercase_ : List[str]=99 , lowercase_ : Union[str, Any]=32 , lowercase_ : List[Any]=5 , lowercase_ : Dict=4 , lowercase_ : List[Any]=37 , lowercase_ : Any="gelu" , lowercase_ : Union[str, Any]=0.1 , lowercase_ : Tuple=0.1 , lowercase_ : List[Any]=512 , lowercase_ : Any=16 , lowercase_ : Optional[int]=2 , lowercase_ : Any=0.02 , lowercase_ : Dict=3 , lowercase_ : Union[str, Any]=4 , lowercase_ : Tuple=None , ) -> List[Any]: """simple docstring""" _lowerCamelCase : List[Any] =parent _lowerCamelCase : Tuple =batch_size _lowerCamelCase : Any =seq_length _lowerCamelCase : int =is_training _lowerCamelCase : int =use_input_mask _lowerCamelCase : List[str] =use_token_type_ids _lowerCamelCase : Dict =use_labels _lowerCamelCase : int =vocab_size _lowerCamelCase : Optional[Any] =hidden_size _lowerCamelCase : Union[str, Any] =num_hidden_layers _lowerCamelCase : Any =num_attention_heads _lowerCamelCase : Tuple =intermediate_size _lowerCamelCase : List[str] =hidden_act _lowerCamelCase : int =hidden_dropout_prob _lowerCamelCase : Optional[int] =attention_probs_dropout_prob _lowerCamelCase : Any =max_position_embeddings _lowerCamelCase : Optional[Any] =type_vocab_size _lowerCamelCase : List[Any] =type_sequence_label_size _lowerCamelCase : Union[str, Any] =initializer_range _lowerCamelCase : Dict =num_labels _lowerCamelCase : Optional[Any] =num_choices _lowerCamelCase : Dict =scope def lowerCamelCase ( self : Dict ) -> int: """simple docstring""" _lowerCamelCase : Optional[int] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : Union[str, Any] =None if self.use_input_mask: _lowerCamelCase : Dict =random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase : str =None if self.use_token_type_ids: _lowerCamelCase : Dict =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : Tuple =None _lowerCamelCase : Optional[Any] =None _lowerCamelCase : Optional[Any] =None if self.use_labels: _lowerCamelCase : int =ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase : List[Any] =ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase : Optional[Any] =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase ( self : int ) -> Tuple: """simple docstring""" return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def lowerCamelCase ( self : Tuple , lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Optional[int]: """simple docstring""" _lowerCamelCase : Dict =LlamaModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : List[str] =model(lowercase_ , attention_mask=lowercase_ ) _lowerCamelCase : List[str] =model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase ( self : int , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : int , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : Any , lowercase_ : str , lowercase_ : Dict , lowercase_ : str , ) -> Any: """simple docstring""" _lowerCamelCase : str =True _lowerCamelCase : str =LlamaModel(lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : int =model( lowercase_ , attention_mask=lowercase_ , encoder_hidden_states=lowercase_ , encoder_attention_mask=lowercase_ , ) _lowerCamelCase : str =model( lowercase_ , attention_mask=lowercase_ , encoder_hidden_states=lowercase_ , ) _lowerCamelCase : List[Any] =model(lowercase_ , attention_mask=lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase ( self : Dict , lowercase_ : Dict , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : Dict , lowercase_ : Tuple , lowercase_ : str , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Any , ) -> Optional[int]: """simple docstring""" _lowerCamelCase : int =LlamaForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : int =model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self : int , lowercase_ : str , lowercase_ : Dict , lowercase_ : str , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : List[Any] , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Optional[Any] , ) -> List[str]: """simple docstring""" _lowerCamelCase : List[str] =True _lowerCamelCase : Tuple =True _lowerCamelCase : str =LlamaForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() # first forward pass _lowerCamelCase : int =model( lowercase_ , attention_mask=lowercase_ , encoder_hidden_states=lowercase_ , encoder_attention_mask=lowercase_ , use_cache=lowercase_ , ) _lowerCamelCase : Any =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _lowerCamelCase : Union[str, Any] =ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCamelCase : Optional[Any] =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _lowerCamelCase : Tuple =torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase : Optional[Any] =torch.cat([input_mask, next_mask] , dim=-1 ) _lowerCamelCase : Optional[int] =model( lowercase_ , attention_mask=lowercase_ , encoder_hidden_states=lowercase_ , encoder_attention_mask=lowercase_ , output_hidden_states=lowercase_ , )['hidden_states'][0] _lowerCamelCase : Tuple =model( lowercase_ , attention_mask=lowercase_ , encoder_hidden_states=lowercase_ , encoder_attention_mask=lowercase_ , past_key_values=lowercase_ , output_hidden_states=lowercase_ , )['hidden_states'][0] # select random slice _lowerCamelCase : int =ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase : Optional[Any] =output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCamelCase : Optional[int] =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase_ , lowercase_ , atol=1E-3 ) ) def lowerCamelCase ( self : Tuple ) -> str: """simple docstring""" _lowerCamelCase : int =self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Optional[Any] =config_and_inputs _lowerCamelCase : Union[str, Any] ={'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): UpperCamelCase__ : Optional[Any] =(LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () UpperCamelCase__ : Union[str, Any] =(LlamaForCausalLM,) if is_torch_available() else () UpperCamelCase__ : List[Any] =( { 'feature-extraction': LlamaModel, 'text-classification': LlamaForSequenceClassification, 'text-generation': LlamaForCausalLM, 'zero-shot': LlamaForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase__ : Optional[Any] =False UpperCamelCase__ : Tuple =False def lowerCamelCase ( self : Union[str, Any] ) -> str: """simple docstring""" _lowerCamelCase : Optional[Any] =LlamaModelTester(self ) _lowerCamelCase : Any =ConfigTester(self , config_class=lowercase_ , hidden_size=37 ) def lowerCamelCase ( self : List[str] ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase ( self : Dict ) -> str: """simple docstring""" _lowerCamelCase : Optional[int] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def lowerCamelCase ( self : int ) -> Tuple: """simple docstring""" _lowerCamelCase : List[str] =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCamelCase : Tuple =type self.model_tester.create_and_check_model(*lowercase_ ) def lowerCamelCase ( self : Tuple ) -> List[str]: """simple docstring""" _lowerCamelCase , _lowerCamelCase : List[str] =self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : str =3 _lowerCamelCase : str =input_dict['input_ids'] _lowerCamelCase : int =input_ids.ne(1 ).to(lowercase_ ) _lowerCamelCase : Optional[Any] =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCamelCase : Union[str, Any] =LlamaForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : Dict =model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCamelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" _lowerCamelCase , _lowerCamelCase : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : Optional[Any] =3 _lowerCamelCase : List[Any] ='single_label_classification' _lowerCamelCase : List[str] =input_dict['input_ids'] _lowerCamelCase : Any =input_ids.ne(1 ).to(lowercase_ ) _lowerCamelCase : List[str] =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCamelCase : str =LlamaForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : Optional[int] =model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCamelCase ( self : Dict ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase : List[str] =self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : int =3 _lowerCamelCase : Optional[Any] ='multi_label_classification' _lowerCamelCase : str =input_dict['input_ids'] _lowerCamelCase : Tuple =input_ids.ne(1 ).to(lowercase_ ) _lowerCamelCase : Tuple =ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _lowerCamelCase : Optional[Any] =LlamaForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() _lowerCamelCase : Optional[Any] =model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def lowerCamelCase ( self : int ) -> str: """simple docstring""" pass @parameterized.expand([('linear',), ('dynamic',)] ) def lowerCamelCase ( self : Optional[int] , lowercase_ : List[Any] ) -> int: """simple docstring""" _lowerCamelCase , _lowerCamelCase : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : Union[str, Any] =ids_tensor([1, 10] , config.vocab_size ) _lowerCamelCase : List[str] =ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCamelCase : Optional[int] =LlamaModel(lowercase_ ) original_model.to(lowercase_ ) original_model.eval() _lowerCamelCase : List[Any] =original_model(lowercase_ ).last_hidden_state _lowerCamelCase : Dict =original_model(lowercase_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCamelCase : Dict ={'type': scaling_type, 'factor': 10.0} _lowerCamelCase : int =LlamaModel(lowercase_ ) scaled_model.to(lowercase_ ) scaled_model.eval() _lowerCamelCase : Optional[int] =scaled_model(lowercase_ ).last_hidden_state _lowerCamelCase : Dict =scaled_model(lowercase_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) @require_torch class A ( unittest.TestCase ): @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCamelCase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase : List[str] =[1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCamelCase : Optional[Any] =LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) _lowerCamelCase : int =model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 _lowerCamelCase : int =torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , lowercase_ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCamelCase : Dict =torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , lowercase_ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCamelCase ( self : int ) -> Any: """simple docstring""" _lowerCamelCase : Tuple =[1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCamelCase : Optional[Any] =LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) _lowerCamelCase : List[Any] =model(torch.tensor(lowercase_ ) ) # Expected mean on dim = -1 _lowerCamelCase : str =torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , lowercase_ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCamelCase : List[str] =torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , lowercase_ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCamelCase ( self : List[Any] ) -> Any: """simple docstring""" _lowerCamelCase : Optional[Any] =[1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCamelCase : str =LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) _lowerCamelCase : int =model(torch.tensor(lowercase_ ) ) # Expected mean on dim = -1 _lowerCamelCase : str =torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , lowercase_ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCamelCase : Union[str, Any] =torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , lowercase_ , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def lowerCamelCase ( self : int ) -> List[Any]: """simple docstring""" _lowerCamelCase : List[str] =[1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCamelCase : Any =LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) _lowerCamelCase : Optional[Any] =model(torch.tensor(lowercase_ ) ) _lowerCamelCase : Optional[int] =torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , lowercase_ , atol=1E-2 , rtol=1E-2 ) # fmt: off _lowerCamelCase : int =torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , lowercase_ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def lowerCamelCase ( self : str ) -> Optional[int]: """simple docstring""" _lowerCamelCase : Tuple ='Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' _lowerCamelCase : Union[str, Any] ='Simply put, the theory of relativity states that ' _lowerCamelCase : int =LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) _lowerCamelCase : str =tokenizer.encode(lowercase_ , return_tensors='pt' ) _lowerCamelCase : List[Any] =LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=lowercase_ ) # greedy generation outputs _lowerCamelCase : str =model.generate(lowercase_ , max_new_tokens=64 , top_p=lowercase_ , temperature=1 , do_sample=lowercase_ ) _lowerCamelCase : Tuple =tokenizer.decode(generated_ids[0] , skip_special_tokens=lowercase_ ) self.assertEqual(lowercase_ , lowercase_ )
199
0
from __future__ import annotations import math def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> int: '''simple docstring''' if depth < 0: raise ValueError('Depth cannot be less than 0' ) if not scores: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , lowercase__ , lowercase__ , lowercase__ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase__ , lowercase__ , lowercase__ ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , lowercase__ , lowercase__ , lowercase__ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase__ , lowercase__ , lowercase__ ) , ) ) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= [9_0, 2_3, 6, 3_3, 2_1, 6_5, 1_2_3, 3_4_4_2_3] __lowercase= math.log(len(lowercase__ ) , 2 ) print(F'Optimal value : {minimax(0 , 0 , lowercase__ , lowercase__ , lowercase__ )}' ) if __name__ == "__main__": import doctest doctest.testmod() main()
356
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent lowerCAmelCase = {'''UserAgent''': UserAgent().random} def _lowerCamelCase( lowercase__ ) -> dict: '''simple docstring''' __lowercase= script.contents[0] __lowercase= json.loads(data[data.find('{"config"' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class A : def __init__(self , lowerCAmelCase ): __lowercase= f'https://www.instagram.com/{username}/' __lowercase= self.get_json() def _A (self ): __lowercase= requests.get(self.url , headers=lowerCAmelCase ).text __lowercase= BeautifulSoup(lowerCAmelCase , 'html.parser' ).find_all('script' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__(self ): return f'{self.__class__.__name__}(\'{self.username}\')' def __str__(self ): return f'{self.fullname} ({self.username}) is {self.biography}' @property def _A (self ): return self.user_data["username"] @property def _A (self ): return self.user_data["full_name"] @property def _A (self ): return self.user_data["biography"] @property def _A (self ): return self.user_data["business_email"] @property def _A (self ): return self.user_data["external_url"] @property def _A (self ): return self.user_data["edge_followed_by"]["count"] @property def _A (self ): return self.user_data["edge_follow"]["count"] @property def _A (self ): return self.user_data["edge_owner_to_timeline_media"]["count"] @property def _A (self ): return self.user_data["profile_pic_url_hd"] @property def _A (self ): return self.user_data["is_verified"] @property def _A (self ): return self.user_data["is_private"] def _lowerCamelCase( lowercase__ = "github" ) -> None: '''simple docstring''' import os if os.environ.get('CI' ): return # test failing on GitHub Actions __lowercase= InstagramUser(lowercase__ ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , lowercase__ ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 1_5_0 assert instagram_user.number_of_followers > 1_2_0_0_0_0 assert instagram_user.number_of_followings > 1_5 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('https://instagram.' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase = InstagramUser('''github''') print(instagram_user) print(F'{instagram_user.number_of_posts = }') print(F'{instagram_user.number_of_followers = }') print(F'{instagram_user.number_of_followings = }') print(F'{instagram_user.email = }') print(F'{instagram_user.website = }') print(F'{instagram_user.profile_picture_url = }') print(F'{instagram_user.is_verified = }') print(F'{instagram_user.is_private = }')
304
0
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html _SCREAMING_SNAKE_CASE = """platform""" import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def SCREAMING_SNAKE_CASE__ ( __a , __a , __a=None , __a=None , __a=None , __a=None , __a=None , __a=None , ): if attention_mask is None: snake_case_ : Optional[int] = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: snake_case_ : int = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: snake_case_ : Any = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ : Any = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case_ : List[str] = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class SCREAMING_SNAKE_CASE_ : def __init__( self : List[Any] , _A : Tuple , _A : Tuple=13 , _A : Union[str, Any]=7 , _A : List[str]=True , _A : Any=False , _A : int=99 , _A : Dict=16 , _A : Tuple=2 , _A : Optional[Any]=4 , _A : Optional[int]=4 , _A : str="gelu" , _A : Dict=0.1 , _A : Tuple=0.1 , _A : Optional[Any]=32 , _A : Tuple=2 , _A : Tuple=1 , _A : Optional[int]=0 , _A : str=0.0_2 , ) -> Tuple: """simple docstring""" snake_case_ : Optional[int] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : List[str] = seq_length snake_case_ : Dict = is_training snake_case_ : Union[str, Any] = use_labels snake_case_ : List[str] = vocab_size snake_case_ : Any = hidden_size snake_case_ : Dict = num_hidden_layers snake_case_ : str = num_attention_heads snake_case_ : str = intermediate_size snake_case_ : Optional[Any] = hidden_act snake_case_ : List[str] = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : Any = eos_token_id snake_case_ : Optional[Any] = pad_token_id snake_case_ : Tuple = bos_token_id snake_case_ : List[Any] = initializer_range def UpperCAmelCase_ ( self : str ) -> Dict: """simple docstring""" snake_case_ : Optional[int] = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) snake_case_ : str = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) snake_case_ : Optional[int] = shift_tokens_right(_UpperCAmelCase , 1 , 2 ) snake_case_ : Optional[int] = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=_UpperCAmelCase , ) snake_case_ : Optional[Any] = prepare_blenderbot_inputs_dict(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return config, inputs_dict def UpperCAmelCase_ ( self : List[str] ) -> Optional[Any]: """simple docstring""" snake_case_ ,snake_case_ : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def UpperCAmelCase_ ( self : int , _A : Optional[int] , _A : List[str] , _A : List[str] ) -> str: """simple docstring""" snake_case_ : Union[str, Any] = 20 snake_case_ : Optional[Any] = model_class_name(_UpperCAmelCase ) snake_case_ : List[Any] = model.encode(inputs_dict['input_ids'] ) snake_case_ ,snake_case_ : Union[str, Any] = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) snake_case_ : Optional[int] = model.init_cache(decoder_input_ids.shape[0] , _UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Tuple = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='i4' ) snake_case_ : Tuple = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) snake_case_ : Optional[Any] = model.decode( decoder_input_ids[:, :-1] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) snake_case_ : Optional[Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) snake_case_ : Dict = model.decode( decoder_input_ids[:, -1:] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=_UpperCAmelCase , ) snake_case_ : Any = model.decode(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Optional[Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) def UpperCAmelCase_ ( self : Union[str, Any] , _A : str , _A : str , _A : List[str] ) -> int: """simple docstring""" snake_case_ : Optional[int] = 20 snake_case_ : Optional[int] = model_class_name(_UpperCAmelCase ) snake_case_ : Union[str, Any] = model.encode(inputs_dict['input_ids'] ) snake_case_ ,snake_case_ : int = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) snake_case_ : Optional[Any] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) snake_case_ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] , _UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Tuple = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) snake_case_ : Optional[int] = model.decode( decoder_input_ids[:, :-1] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) snake_case_ : Dict = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) snake_case_ : str = model.decode( decoder_input_ids[:, -1:] , _UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) snake_case_ : Union[str, Any] = model.decode(_UpperCAmelCase , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase ) snake_case_ : Dict = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) @require_flax class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): __magic_name__: Any = 99 def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" snake_case_ : Optional[Any] = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) snake_case_ : Union[str, Any] = input_ids.shape[0] snake_case_ : Optional[int] = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def UpperCAmelCase_ ( self : int ) -> Any: """simple docstring""" snake_case_ ,snake_case_ ,snake_case_ : Optional[int] = self._get_config_and_data() snake_case_ : Union[str, Any] = FlaxBlenderbotForConditionalGeneration(_UpperCAmelCase ) snake_case_ : str = lm_model(input_ids=_UpperCAmelCase ) snake_case_ : Any = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs['logits'].shape , _UpperCAmelCase ) def UpperCAmelCase_ ( self : Any ) -> Dict: """simple docstring""" snake_case_ : int = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) snake_case_ : Optional[Any] = FlaxBlenderbotForConditionalGeneration(_UpperCAmelCase ) snake_case_ : int = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) snake_case_ : str = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) snake_case_ : Tuple = lm_model(input_ids=_UpperCAmelCase , decoder_input_ids=_UpperCAmelCase ) snake_case_ : Optional[Any] = (*summary.shape, config.vocab_size) self.assertEqual(outputs['logits'].shape , _UpperCAmelCase ) def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: """simple docstring""" snake_case_ : str = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) snake_case_ : Union[str, Any] = shift_tokens_right(_UpperCAmelCase , 1 , 2 ) snake_case_ : Optional[Any] = np.equal(_UpperCAmelCase , 1 ).astype(np.floataa ).sum() snake_case_ : List[str] = np.equal(_UpperCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(_UpperCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class SCREAMING_SNAKE_CASE_ ( lowerCAmelCase__ , unittest.TestCase , lowerCAmelCase__ ): __magic_name__: Any = True __magic_name__: Tuple = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) __magic_name__: str = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" snake_case_ : int = FlaxBlenderbotModelTester(self ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[Any]: """simple docstring""" snake_case_ ,snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase_ ( self : List[Any] ) -> int: """simple docstring""" snake_case_ ,snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase_ ( self : Dict ) -> List[Any]: """simple docstring""" snake_case_ ,snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): snake_case_ : List[Any] = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Tuple = model_class(_UpperCAmelCase ) @jax.jit def encode_jitted(_A : List[Any] , _A : List[Any]=None , **_A : str ): return model.encode(input_ids=_UpperCAmelCase , attention_mask=_UpperCAmelCase ) with self.subTest('JIT Enabled' ): snake_case_ : Dict = encode_jitted(**_UpperCAmelCase ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): snake_case_ : str = encode_jitted(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) ) for jitted_output, output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" snake_case_ ,snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): snake_case_ : List[Any] = model_class(_UpperCAmelCase ) snake_case_ : Tuple = model.encode(inputs_dict['input_ids'] , inputs_dict['attention_mask'] ) snake_case_ : Optional[Any] = { 'decoder_input_ids': inputs_dict['decoder_input_ids'], 'decoder_attention_mask': inputs_dict['decoder_attention_mask'], 'encoder_outputs': encoder_outputs, } @jax.jit def decode_jitted(_A : int , _A : List[Any] , _A : int ): return model.decode( decoder_input_ids=_UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , encoder_outputs=_UpperCAmelCase , ) with self.subTest('JIT Enabled' ): snake_case_ : str = decode_jitted(**_UpperCAmelCase ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): snake_case_ : Dict = decode_jitted(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) ) for jitted_output, output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCAmelCase_ ( self : Dict ) -> Optional[Any]: """simple docstring""" for model_class_name in self.all_model_classes: snake_case_ : Union[str, Any] = model_class_name.from_pretrained('facebook/blenderbot-400M-distill' ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids snake_case_ : Union[str, Any] = np.ones((1, 1) ) * model.config.eos_token_id snake_case_ : List[str] = model(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) @unittest.skipUnless(jax_device != 'cpu' , '3B test too slow on CPU.' ) @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" snake_case_ : List[Any] = {'num_beams': 1, 'early_stopping': True, 'min_length': 15, 'max_length': 25} snake_case_ : int = {'skip_special_tokens': True, 'clean_up_tokenization_spaces': True} snake_case_ : Optional[int] = FlaxBlenderbotForConditionalGeneration.from_pretrained('facebook/blenderbot-3B' , from_pt=_UpperCAmelCase ) snake_case_ : Optional[int] = BlenderbotTokenizer.from_pretrained('facebook/blenderbot-3B' ) snake_case_ : Dict = ['Sam'] snake_case_ : str = tokenizer(_UpperCAmelCase , return_tensors='jax' ) snake_case_ : str = model.generate(**_UpperCAmelCase , **_UpperCAmelCase ) snake_case_ : Optional[int] = 'Sam is a great name. It means "sun" in Gaelic.' snake_case_ : Optional[Any] = tokenizer.batch_decode(_UpperCAmelCase , **_UpperCAmelCase ) assert generated_txt[0].strip() == tgt_text
327
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class A__ ( lowerCAmelCase__ ): lowerCAmelCase__ : Union[str, Any] = "transfo-xl" lowerCAmelCase__ : int = ["mems"] lowerCAmelCase__ : Dict = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , _UpperCAmelCase : Tuple=26_77_35 , _UpperCAmelCase : Any=[2_00_00, 4_00_00, 20_00_00] , _UpperCAmelCase : Tuple=10_24 , _UpperCAmelCase : Union[str, Any]=10_24 , _UpperCAmelCase : Optional[int]=16 , _UpperCAmelCase : Tuple=64 , _UpperCAmelCase : Tuple=40_96 , _UpperCAmelCase : List[Any]=4 , _UpperCAmelCase : str=False , _UpperCAmelCase : Optional[Any]=18 , _UpperCAmelCase : int=16_00 , _UpperCAmelCase : Optional[int]=10_00 , _UpperCAmelCase : Optional[Any]=True , _UpperCAmelCase : Tuple=True , _UpperCAmelCase : Any=0 , _UpperCAmelCase : Optional[Any]=-1 , _UpperCAmelCase : List[str]=True , _UpperCAmelCase : Optional[Any]=0.1 , _UpperCAmelCase : List[str]=0.0 , _UpperCAmelCase : Optional[Any]=True , _UpperCAmelCase : int="normal" , _UpperCAmelCase : int=0.01 , _UpperCAmelCase : List[Any]=0.01 , _UpperCAmelCase : List[Any]=0.02 , _UpperCAmelCase : Optional[Any]=1e-5 , _UpperCAmelCase : Tuple=0 , **_UpperCAmelCase : List[str] , ) -> Tuple: """simple docstring""" __lowercase = vocab_size __lowercase = [] self.cutoffs.extend(_UpperCAmelCase ) if proj_share_all_but_first: __lowercase = [False] + [True] * len(self.cutoffs ) else: __lowercase = [False] + [False] * len(self.cutoffs ) __lowercase = d_model __lowercase = d_embed __lowercase = d_head __lowercase = d_inner __lowercase = div_val __lowercase = pre_lnorm __lowercase = n_layer __lowercase = n_head __lowercase = mem_len __lowercase = same_length __lowercase = attn_type __lowercase = clamp_len __lowercase = sample_softmax __lowercase = adaptive __lowercase = dropout __lowercase = dropatt __lowercase = untie_r __lowercase = init __lowercase = init_range __lowercase = proj_init_std __lowercase = init_std __lowercase = layer_norm_epsilon super().__init__(eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) @property def a__ ( self : Tuple ) -> Any: """simple docstring""" logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def a__ ( self : Dict , _UpperCAmelCase : List[str] ) -> Optional[Any]: """simple docstring""" raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
325
0
from functools import reduce lowercase = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def lowerCamelCase_ ( UpperCamelCase__ : str = N ): '''simple docstring''' return max( # mypy cannot properly interpret reduce int(reduce(lambda UpperCamelCase__, UpperCamelCase__ : str(int(UpperCamelCase__ ) * int(UpperCamelCase__ ) ), n[i : i + 13] ) ) for i in range(len(UpperCamelCase__ ) - 12 ) ) if __name__ == "__main__": print(f'{solution() = }')
35
from __future__ import annotations from collections import Counter from random import random class __lowercase : '''simple docstring''' def __init__( self : List[Any] ): UpperCamelCase__ = {} def A_ ( self : List[Any] , _a : str ): UpperCamelCase__ = {} def A_ ( self : List[Any] , _a : str , _a : str , _a : float ): if nodea not in self.connections: self.add_node(_a ) if nodea not in self.connections: self.add_node(_a ) UpperCamelCase__ = probability def A_ ( self : Optional[Any] ): return list(self.connections ) def A_ ( self : Tuple , _a : str ): UpperCamelCase__ = 0 UpperCamelCase__ = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def lowerCamelCase_ ( UpperCamelCase__ : str, UpperCamelCase__ : list[tuple[str, str, float]], UpperCamelCase__ : int ): '''simple docstring''' UpperCamelCase__ = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ) UpperCamelCase__ = Counter(graph.get_nodes() ) UpperCamelCase__ = start for _ in range(UpperCamelCase__ ): UpperCamelCase__ = graph.transition(UpperCamelCase__ ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
35
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer __snake_case : Optional[int] = logging.get_logger(__name__) __snake_case : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} __snake_case : Optional[Any] = { 'vocab_file': { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt' ), } } __snake_case : Tuple = { 'junnyu/roformer_chinese_small': 1536, 'junnyu/roformer_chinese_base': 1536, 'junnyu/roformer_chinese_char_small': 512, 'junnyu/roformer_chinese_char_base': 512, 'junnyu/roformer_small_discriminator': 128, 'junnyu/roformer_small_generator': 128, } __snake_case : Optional[Any] = { 'junnyu/roformer_chinese_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_base': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_base': {'do_lower_case': True}, 'junnyu/roformer_small_discriminator': {'do_lower_case': True}, 'junnyu/roformer_small_generator': {'do_lower_case': True}, } class lowerCamelCase ( lowercase_ ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = PRETRAINED_INIT_CONFIGURATION __snake_case = RoFormerTokenizer def __init__( self : str , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any="[UNK]" , lowerCAmelCase_ : List[Any]="[SEP]" , lowerCAmelCase_ : Union[str, Any]="[PAD]" , lowerCAmelCase_ : Optional[Any]="[CLS]" , lowerCAmelCase_ : Dict="[MASK]" , lowerCAmelCase_ : Optional[int]=True , lowerCAmelCase_ : Tuple=None , **lowerCAmelCase_ : Tuple , ) -> List[str]: '''simple docstring''' super().__init__( lowerCAmelCase_ , tokenizer_file=lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , tokenize_chinese_chars=lowerCAmelCase_ , strip_accents=lowerCAmelCase_ , **lowerCAmelCase_ , ) A__ : Union[str, Any] =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get("""lowercase""" , lowerCAmelCase_ ) != do_lower_case or pre_tok_state.get("""strip_accents""" , lowerCAmelCase_ ) != strip_accents ): A__ : int =getattr(lowerCAmelCase_ , pre_tok_state.pop("""type""" ) ) A__ : Union[str, Any] =do_lower_case A__ : Tuple =strip_accents A__ : int =pre_tok_class(**lowerCAmelCase_ ) A__ : List[Any] =do_lower_case def __getstate__( self : Optional[int] ) -> str: '''simple docstring''' A__ : Any =self.__dict__.copy() A__ : List[str] =BertPreTokenizer() return state def __setstate__( self : int , lowerCAmelCase_ : str ) -> str: '''simple docstring''' A__ : str =d A__ : Optional[Any] =self.__dict__["""_tokenizer"""].get_vocab() A__ : Any =PreTokenizer.custom(JiebaPreTokenizer(lowerCAmelCase_ ) ) def lowercase__ ( self : int , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=None ) -> Optional[Any]: '''simple docstring''' A__ : List[str] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase__ ( self : Optional[Any] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' A__ : int =[self.sep_token_id] A__ : List[str] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase__ ( self : List[str] , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' A__ : List[Any] =self._tokenizer.model.save(lowerCAmelCase_ , name=lowerCAmelCase_ ) return tuple(lowerCAmelCase_ ) def lowercase__ ( self : Any , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : str=None , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : Tuple=False , **lowerCAmelCase_ : Tuple , ) -> List[Any]: '''simple docstring''' A__ : List[Any] =BertPreTokenizer() return super().save_pretrained(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ )
134
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __snake_case : Any = pytest.mark.integration @pytest.mark.parametrize("""path""", ["""paws""", """csv"""] ) def __lowerCamelCase ( __snake_case : Any, __snake_case : List[Any] ) -> Optional[int]: """simple docstring""" inspect_dataset(__snake_case, __snake_case ) A__ : Dict =path + """.py""" assert script_name in os.listdir(__snake_case ) assert "__pycache__" not in os.listdir(__snake_case ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""", ["""accuracy"""] ) def __lowerCamelCase ( __snake_case : Tuple, __snake_case : Any ) -> int: """simple docstring""" inspect_metric(__snake_case, __snake_case ) A__ : int =path + """.py""" assert script_name in os.listdir(__snake_case ) assert "__pycache__" not in os.listdir(__snake_case ) @pytest.mark.parametrize( """path, config_name, expected_splits""", [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ], ) def __lowerCamelCase ( __snake_case : List[str], __snake_case : Union[str, Any], __snake_case : Optional[int] ) -> Optional[int]: """simple docstring""" A__ : Optional[Any] =get_dataset_config_info(__snake_case, config_name=__snake_case ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""", [ ("""paws""", None, ValueError), ], ) def __lowerCamelCase ( __snake_case : Optional[int], __snake_case : Optional[Any], __snake_case : List[Any] ) -> List[str]: """simple docstring""" with pytest.raises(__snake_case ): get_dataset_config_info(__snake_case, config_name=__snake_case ) @pytest.mark.parametrize( """path, expected""", [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ], ) def __lowerCamelCase ( __snake_case : Tuple, __snake_case : str ) -> Optional[Any]: """simple docstring""" A__ : Optional[int] =get_dataset_config_names(__snake_case ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""", [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ], ) def __lowerCamelCase ( __snake_case : List[Any], __snake_case : Dict, __snake_case : Optional[Any] ) -> int: """simple docstring""" A__ : List[Any] =get_dataset_infos(__snake_case ) assert list(infos.keys() ) == expected_configs A__ : Optional[int] =expected_configs[0] assert expected_config in infos A__ : List[str] =infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""", [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ], ) def __lowerCamelCase ( __snake_case : Optional[int], __snake_case : List[Any], __snake_case : Optional[Any] ) -> Any: """simple docstring""" A__ : str =get_dataset_infos(__snake_case ) assert expected_config in infos A__ : int =infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""", [ ("""paws""", None, ValueError), ], ) def __lowerCamelCase ( __snake_case : Dict, __snake_case : Optional[int], __snake_case : Dict ) -> Tuple: """simple docstring""" with pytest.raises(__snake_case ): get_dataset_split_names(__snake_case, config_name=__snake_case )
134
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Tuple=7 , lowerCAmelCase__ : List[Any]=3 , lowerCAmelCase__ : Optional[int]=18 , lowerCAmelCase__ : Any=30 , lowerCAmelCase__ : int=400 , lowerCAmelCase__ : List[str]=True , lowerCAmelCase__ : Optional[Any]=None , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : int=False , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : int=True , lowerCAmelCase__ : Dict=[0.5, 0.5, 0.5] , lowerCAmelCase__ : str=[0.5, 0.5, 0.5] , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size if size is not None else {'''height''': 18, '''width''': 20} _UpperCamelCase = do_thumbnail _UpperCamelCase = do_align_axis _UpperCamelCase = do_pad _UpperCamelCase = do_normalize _UpperCamelCase = image_mean _UpperCamelCase = image_std def snake_case__ ( self : List[Any] ) -> Any: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowerCAmelCase ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" _snake_case : Tuple = DonutImageProcessor if is_vision_available() else None def snake_case__ ( self : str ) -> int: '''simple docstring''' _UpperCamelCase = DonutImageProcessingTester(self ) @property def snake_case__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self : Any ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''size''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_thumbnail''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_align_long_axis''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_pad''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_mean''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_std''' ) ) def snake_case__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 20} ) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) # Previous config had dimensions in (width, height) order _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {'''height''': 84, '''width''': 42} ) def snake_case__ ( self : Any ) -> str: '''simple docstring''' pass @is_flaky() def snake_case__ ( self : Optional[int] ) -> Any: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) @is_flaky() def snake_case__ ( self : Optional[int] ) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) @is_flaky() def snake_case__ ( self : str ) -> Tuple: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , )
357
'''simple docstring''' import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowercase__ : str = 16 lowercase__ : int = 32 def a__ ( lowercase : Accelerator, lowercase : int = 16 ) -> List[str]: """simple docstring""" _UpperCamelCase = AutoTokenizer.from_pretrained('''bert-base-cased''' ) _UpperCamelCase = load_dataset('''glue''', '''mrpc''' ) def tokenize_function(lowercase : Union[str, Any] ): # max_length=None => use the model max length (it's actually the default) _UpperCamelCase = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=lowercase, max_length=lowercase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _UpperCamelCase = datasets.map( lowercase, batched=lowercase, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _UpperCamelCase = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(lowercase : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. _UpperCamelCase = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _UpperCamelCase = 16 elif accelerator.mixed_precision != "no": _UpperCamelCase = 8 else: _UpperCamelCase = None return tokenizer.pad( lowercase, padding='''longest''', max_length=lowercase, pad_to_multiple_of=lowercase, return_tensors='''pt''', ) # Instantiate dataloaders. _UpperCamelCase = DataLoader( tokenized_datasets['''train'''], shuffle=lowercase, collate_fn=lowercase, batch_size=lowercase ) _UpperCamelCase = DataLoader( tokenized_datasets['''validation'''], shuffle=lowercase, collate_fn=lowercase, batch_size=lowercase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowercase__ : str = mocked_dataloaders # noqa: F811 def a__ ( lowercase : List[Any], lowercase : List[str] ) -> Any: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''', lowercase ) == "1": _UpperCamelCase = 2 # Initialize accelerator _UpperCamelCase = Accelerator(cpu=args.cpu, mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _UpperCamelCase = config['''lr'''] _UpperCamelCase = int(config['''num_epochs'''] ) _UpperCamelCase = int(config['''seed'''] ) _UpperCamelCase = int(config['''batch_size'''] ) _UpperCamelCase = evaluate.load('''glue''', '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=lowercase ) def inner_training_loop(lowercase : List[Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(lowercase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _UpperCamelCase = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''', return_dict=lowercase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _UpperCamelCase = model.to(accelerator.device ) # Instantiate optimizer _UpperCamelCase = AdamW(params=model.parameters(), lr=lowercase ) _UpperCamelCase , _UpperCamelCase = get_dataloaders(lowercase, lowercase ) # Instantiate scheduler _UpperCamelCase = get_linear_schedule_with_warmup( optimizer=lowercase, num_warmup_steps=100, num_training_steps=(len(lowercase ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = accelerator.prepare( lowercase, lowercase, lowercase, lowercase, lowercase ) # Now we train the model for epoch in range(lowercase ): model.train() for step, batch in enumerate(lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _UpperCamelCase = model(**lowercase ) _UpperCamelCase = outputs.loss accelerator.backward(lowercase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _UpperCamelCase = model(**lowercase ) _UpperCamelCase = outputs.logits.argmax(dim=-1 ) _UpperCamelCase , _UpperCamelCase = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=lowercase, references=lowercase, ) _UpperCamelCase = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""", lowercase ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a__ ( ) -> str: """simple docstring""" _UpperCamelCase = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''', type=lowercase, default=lowercase, choices=['''no''', '''fp16''', '''bf16''', '''fp8'''], help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''', ) parser.add_argument('''--cpu''', action='''store_true''', help='''If passed, will train on the CPU.''' ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(lowercase, lowercase ) if __name__ == "__main__": main()
287
0
"""simple docstring""" class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [ [], [], [], ] def snake_case ( self , __a , __a ): try: if len(self.queues[priority] ) >= 1_00: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(__a ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def snake_case ( self ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self ): return "\n".join(f"Priority {i}: {q}" for i, q in enumerate(self.queues ) ) class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [] def snake_case ( self , __a ): if len(self.queue ) == 1_00: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(__a ) def snake_case ( self ): if not self.queue: raise UnderFlowError("The queue is empty" ) else: __lowerCAmelCase = min(self.queue ) self.queue.remove(__a ) return data def __str__( self ): return str(self.queue ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
57
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] =["""image_processor""", """tokenizer"""] __UpperCAmelCase : Optional[Any] ="""CLIPImageProcessor""" __UpperCAmelCase : Union[str, Any] =("""XLMRobertaTokenizer""", """XLMRobertaTokenizerFast""") def __init__( self , __a=None , __a=None , **__a ): __lowerCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __a , ) __lowerCAmelCase = kwargs.pop("feature_extractor" ) __lowerCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__a , __a ) def __call__( self , __a=None , __a=None , __a=None , **__a ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: __lowerCAmelCase = self.tokenizer(__a , return_tensors=__a , **__a ) if images is not None: __lowerCAmelCase = self.image_processor(__a , return_tensors=__a , **__a ) if text is not None and images is not None: __lowerCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__a ) , tensor_type=__a ) def snake_case ( self , *__a , **__a ): return self.tokenizer.batch_decode(*__a , **__a ) def snake_case ( self , *__a , **__a ): return self.tokenizer.decode(*__a , **__a ) @property def snake_case ( self ): __lowerCAmelCase = self.tokenizer.model_input_names __lowerCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
57
1
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class lowercase_ : '''simple docstring''' def __init__( self : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int]=13 , __UpperCAmelCase : Optional[int]=7 , __UpperCAmelCase : Optional[int]=True , __UpperCAmelCase : List[str]=True , __UpperCAmelCase : Any=True , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Union[str, Any]=99 , __UpperCAmelCase : int=64 , __UpperCAmelCase : Any=32 , __UpperCAmelCase : Union[str, Any]=5 , __UpperCAmelCase : Optional[Any]=4 , __UpperCAmelCase : Optional[Any]=37 , __UpperCAmelCase : Dict="gelu" , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : List[str]=0.1 , __UpperCAmelCase : Union[str, Any]=512 , __UpperCAmelCase : Union[str, Any]=16 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : List[Any]=0.02 , __UpperCAmelCase : List[Any]=3 , __UpperCAmelCase : Union[str, Any]=4 , __UpperCAmelCase : int=None , ) ->List[str]: """simple docstring""" a = parent a = batch_size a = seq_length a = is_training a = use_input_mask a = use_token_type_ids a = use_labels a = vocab_size a = hidden_size a = embedding_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = type_sequence_label_size a = initializer_range a = num_labels a = num_choices a = scope def __lowerCAmelCase ( self : Dict ) ->List[str]: """simple docstring""" a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a = None if self.use_input_mask: a = random_attention_mask([self.batch_size, self.seq_length] ) a = None if self.use_token_type_ids: a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a = None a = None a = None if self.use_labels: a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a = ids_tensor([self.batch_size] , self.num_choices ) a = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self : int ) ->Dict: """simple docstring""" return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self : List[str] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : str , __UpperCAmelCase : Any , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[str] ) ->List[Any]: """simple docstring""" a = MegatronBertModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase ) a = model(__UpperCAmelCase , token_type_ids=__UpperCAmelCase ) a = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __lowerCAmelCase ( self : Dict , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : List[str] ) ->str: """simple docstring""" a = MegatronBertForMaskedLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Dict , __UpperCAmelCase : List[str] ) ->Optional[int]: """simple docstring""" a = MegatronBertForCausalLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : Optional[int] ) ->Dict: """simple docstring""" a = MegatronBertForNextSentencePrediction(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def __lowerCAmelCase ( self : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : str ) ->List[str]: """simple docstring""" a = MegatronBertForPreTraining(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , next_sentence_label=__UpperCAmelCase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Dict , __UpperCAmelCase : Any , __UpperCAmelCase : List[Any] ) ->Union[str, Any]: """simple docstring""" a = MegatronBertForQuestionAnswering(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , start_positions=__UpperCAmelCase , end_positions=__UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : Any , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) ->Optional[Any]: """simple docstring""" a = self.num_labels a = MegatronBertForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : Any , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple ) ->Dict: """simple docstring""" a = self.num_labels a = MegatronBertForTokenClassification(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str ) ->Optional[int]: """simple docstring""" a = self.num_choices a = MegatronBertForMultipleChoice(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() a = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() a = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() a = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self : Dict ) ->Any: """simple docstring""" a = self.prepare_config_and_inputs() ( ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ) = config_and_inputs a = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase_ ( lowercase , lowercase , unittest.TestCase ): '''simple docstring''' __snake_case = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) __snake_case = ( { '''feature-extraction''': MegatronBertModel, '''fill-mask''': MegatronBertForMaskedLM, '''question-answering''': MegatronBertForQuestionAnswering, '''text-classification''': MegatronBertForSequenceClassification, '''text-generation''': MegatronBertForCausalLM, '''token-classification''': MegatronBertForTokenClassification, '''zero-shot''': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case = True # test_resize_embeddings = False __snake_case = False def __lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : int , __UpperCAmelCase : List[str]=False ) ->Union[str, Any]: """simple docstring""" a = super()._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase , return_labels=__UpperCAmelCase ) if return_labels: if model_class in get_values(__UpperCAmelCase ): a = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__UpperCAmelCase ) a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__UpperCAmelCase ) return inputs_dict def __lowerCAmelCase ( self : Optional[Any] ) ->Any: """simple docstring""" a = MegatronBertModelTester(self ) a = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37 ) def __lowerCAmelCase ( self : List[Any] ) ->Dict: """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self : List[str] ) ->Tuple: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Optional[Any]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Optional[int] ) ->Optional[int]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*__UpperCAmelCase ) def __lowerCAmelCase ( self : List[str] ) ->Optional[int]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Optional[Any]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Optional[Any] ) ->Union[str, Any]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Tuple ) ->Optional[Any]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*__UpperCAmelCase ) def __lowerCAmelCase ( self : Union[str, Any] ) ->List[str]: """simple docstring""" a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*__UpperCAmelCase ) def _a ( a :Union[str, Any] ) -> int: return torch.tensor( a , dtype=torch.long , device=a , ) UpperCAmelCase__ = 1E-4 @require_torch @require_sentencepiece @require_tokenizers class lowercase_ ( unittest.TestCase ): '''simple docstring''' @slow @unittest.skip('''Model is not available.''' ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Dict: """simple docstring""" a = '''nvidia/megatron-bert-uncased-345m''' if "MYDIR" in os.environ: a = os.path.join(os.environ['''MYDIR'''] , __UpperCAmelCase ) a = MegatronBertModel.from_pretrained(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.half() a = _long_tensor([[101, 7_110, 1_005, 1_056, 2_023, 11_333, 17_413, 1_029, 102]] ) with torch.no_grad(): a = model(__UpperCAmelCase )[0] a = torch.Size((1, 9, 1_024) ) self.assertEqual(output.shape , __UpperCAmelCase ) a = [-0.6040, -0.2517, -0.1025, 0.3420, -0.6758, -0.0017, -0.1089, -0.1990, 0.5728] for ii in range(3 ): for jj in range(3 ): a = output[0, ii, jj] a = expected[3 * ii + jj] a = '''ii={} jj={} a={} b={}'''.format(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) self.assertTrue(math.isclose(__UpperCAmelCase , __UpperCAmelCase , rel_tol=__UpperCAmelCase , abs_tol=__UpperCAmelCase ) , msg=__UpperCAmelCase )
350
def _a ( a :int = 600_851_475_143 ) -> int: try: a = int(a ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) a = 2 a = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 a = i while n % i == 0: a = n // i i += 1 return int(a ) if __name__ == "__main__": print(f"""{solution() = }""")
26
0
import math import random from typing import Any from .hill_climbing import SearchProblem def lowerCamelCase_ ( UpperCamelCase__ : Tuple , UpperCamelCase__ : bool = True , UpperCamelCase__ : float = math.inf , UpperCamelCase__ : float = -math.inf , UpperCamelCase__ : float = math.inf , UpperCamelCase__ : float = -math.inf , UpperCamelCase__ : bool = False , UpperCamelCase__ : float = 100 , UpperCamelCase__ : float = 0.01 , UpperCamelCase__ : float = 1 , ) -> Any: """simple docstring""" __lowerCamelCase = False __lowerCamelCase = search_prob __lowerCamelCase = start_temperate __lowerCamelCase = [] __lowerCamelCase = 0 __lowerCamelCase = None while not search_end: __lowerCamelCase = current_state.score() if best_state is None or current_score > best_state.score(): __lowerCamelCase = current_state scores.append(UpperCamelCase__ ) iterations += 1 __lowerCamelCase = None __lowerCamelCase = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to __lowerCamelCase = random.randint(0 , len(UpperCamelCase__ ) - 1 ) # picking a random neighbor __lowerCamelCase = neighbors.pop(UpperCamelCase__ ) __lowerCamelCase = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: __lowerCamelCase = change * -1 # in case we are finding minimum if change > 0: # improves the solution __lowerCamelCase = picked_neighbor else: __lowerCamelCase = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability __lowerCamelCase = picked_neighbor __lowerCamelCase = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor __lowerCamelCase = True else: __lowerCamelCase = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(UpperCamelCase__ ) , UpperCamelCase__ ) plt.xlabel('Iterations' ) plt.ylabel('Function values' ) plt.show() return best_state if __name__ == "__main__": def lowerCamelCase_ ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Tuple ) -> Tuple: """simple docstring""" return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) __A = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __A = simulated_annealing( prob, find_max=False, max_x=1_00, min_x=5, max_y=50, min_y=-5, visualization=True ) print( "The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 " f'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) # starting the problem with initial coordinates (12, 47) __A = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __A = simulated_annealing( prob, find_max=True, max_x=1_00, min_x=5, max_y=50, min_y=-5, visualization=True ) print( "The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 " f'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) def lowerCamelCase_ ( UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple ) -> List[Any]: """simple docstring""" return (3 * x**2) - (6 * y) __A = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __A = simulated_annealing(prob, find_max=False, visualization=True) print( "The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: " f'''{local_min.score()}''' ) __A = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __A = simulated_annealing(prob, find_max=True, visualization=True) print( "The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: " f'''{local_min.score()}''' )
90
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase_ = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __A( __lowerCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = DebertaVaTokenizer SCREAMING_SNAKE_CASE__ = DebertaVaTokenizerFast SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = True def UpperCAmelCase_ (self ): super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , unk_token="""<unk>""" ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = """this is a test""" UpperCamelCase__ = """this is a test""" return input_text, output_text def UpperCAmelCase_ (self ): UpperCamelCase__ = """<pad>""" UpperCamelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<pad>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """[PAD]""" ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 3_00_01 ) def UpperCAmelCase_ (self ): self.assertEqual(self.get_tokenizer().vocab_size , 3_00_00 ) def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """ \tHeLLo!how \n Are yoU? """ UpperCamelCase__ = ["""▁hello""", """!""", """how""", """▁are""", """▁you""", """?"""] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) @unittest.skip("""There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.""" ) def UpperCAmelCase_ (self ): pass @unittest.skip("""There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.""" ) def UpperCAmelCase_ (self ): pass def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = ["""▁""", """<unk>""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """▁""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """<unk>""", """▁""", """.""", ] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = ["""▁i""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """▁""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """<unk>""", """▁""", """.""", ] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = ["""▁i""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """<unk>""", """.""", ] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = ["""▁""", """<unk>""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """▁""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """<unk>""", """▁""", """.""", ] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = """ \tHeLLo!how \n Are yoU? """ UpperCamelCase__ = ["""▁""", """<unk>""", """e""", """<unk>""", """o""", """!""", """how""", """▁""", """<unk>""", """re""", """▁yo""", """<unk>""", """?"""] # fmt: on UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ , split_by_punct=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = self.get_tokenizer() UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = """This is a test""" UpperCamelCase__ = [13, 1, 43_98, 25, 21, 12_89] UpperCamelCase__ = ["""▁""", """T""", """his""", """▁is""", """▁a""", """▁test"""] UpperCamelCase__ = ["""▁""", """<unk>""", """his""", """▁is""", """▁a""", """▁test"""] UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # fmt: off UpperCamelCase__ = """I was born in 92000, and this is falsé.""" UpperCamelCase__ = [13, 1, 23, 3_86, 19, 5_61, 30_50, 15, 17, 48, 25, 82_56, 18, 1, 9] UpperCamelCase__ = ["""▁""", """I""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """é""", """.""", ] UpperCamelCase__ = ["""▁""", """<unk>""", """▁was""", """▁born""", """▁in""", """▁9""", """2000""", """,""", """▁and""", """▁this""", """▁is""", """▁fal""", """s""", """<unk>""", """.""", ] # fmt: on UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.encode("""sequence builders""" ) UpperCamelCase__ = tokenizer.encode("""multi-sequence build""" ) UpperCamelCase__ = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE_ ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE_ , ) @slow def UpperCAmelCase_ (self ): # fmt: off UpperCamelCase__ = {"""input_ids""": [[1, 3_98_67, 36, 1_93_90, 4_86, 27, 3_50_52, 8_14_36, 18, 6_06_85, 12_25, 7, 3_50_52, 8_14_36, 18, 93_67, 1_68_99, 18, 1_59_37, 53, 5_94, 7_73, 18, 1_62_87, 3_04_65, 36, 1_59_37, 6, 4_11_39, 38, 3_69_79, 6_07_63, 1_91, 6, 3_41_32, 99, 6, 5_05_38, 3_90, 4_32_30, 6, 3_41_32, 27_79, 2_08_50, 14, 6_99, 10_72, 11_94, 36, 3_82, 1_09_01, 53, 7, 6_99, 10_72, 20_84, 36, 2_04_22, 6_30, 53, 19, 1_05, 30_49, 18_96, 10_53, 1_68_99, 15_06, 11, 3_79_78, 42_43, 7, 12_37, 3_18_69, 2_00, 1_65_66, 6_54, 6, 3_50_52, 8_14_36, 7, 5_56_30, 1_35_93, 4, 2], [1, 26, 1_50_11, 13, 6_67, 8, 10_53, 18, 2_36_11, 12_37, 7_23_56, 1_28_20, 34, 10_41_34, 12_09, 35, 1_33_13, 66_27, 21, 2_02, 3_47, 7, 1_64, 23_99, 11, 46, 44_85, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 12_32, 28_64, 1_57_85, 1_49_51, 1_05, 5, 85_81, 12_50, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=SCREAMING_SNAKE_CASE_ , model_name="""microsoft/deberta-v2-xlarge""" , revision="""ad6e42c1532ddf3a15c39246b63f5559d558b670""" , )
244
0
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : Optional[Any] = re.compile(R'''\b(a|an|the)\b''', re.UNICODE) snake_case : int = None def __lowerCamelCase ( ): """simple docstring""" a :Optional[Any] = argparse.ArgumentParser('''Official evaluation script for SQuAD version 2.0.''' ) parser.add_argument('''data_file''' , metavar='''data.json''' , help='''Input data JSON file.''' ) parser.add_argument('''pred_file''' , metavar='''pred.json''' , help='''Model predictions.''' ) parser.add_argument( '''--out-file''' , '''-o''' , metavar='''eval.json''' , help='''Write accuracy metrics to file (default is stdout).''' ) parser.add_argument( '''--na-prob-file''' , '''-n''' , metavar='''na_prob.json''' , help='''Model estimates of probability of no answer.''' ) parser.add_argument( '''--na-prob-thresh''' , '''-t''' , type=UpperCAmelCase_ , default=1.0 , help='''Predict "" if no-answer probability exceeds this (default = 1.0).''' , ) parser.add_argument( '''--out-image-dir''' , '''-p''' , metavar='''out_images''' , default=UpperCAmelCase_ , help='''Save precision-recall curves to directory.''' ) parser.add_argument('''--verbose''' , '''-v''' , action='''store_true''' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" a :Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a :str = bool(qa['''answers''']['''text'''] ) return qid_to_has_ans def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" def remove_articles(UpperCAmelCase_ : Tuple ): return ARTICLES_REGEX.sub(''' ''' , UpperCAmelCase_ ) def white_space_fix(UpperCAmelCase_ : int ): return " ".join(text.split() ) def remove_punc(UpperCAmelCase_ : Dict ): a :Optional[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(UpperCAmelCase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(UpperCAmelCase_ ) ) ) ) def __lowerCamelCase ( UpperCAmelCase_ : List[Any] ): """simple docstring""" if not s: return [] return normalize_answer(UpperCAmelCase_ ).split() def __lowerCamelCase ( UpperCAmelCase_ : Any , UpperCAmelCase_ : Any ): """simple docstring""" return int(normalize_answer(UpperCAmelCase_ ) == normalize_answer(UpperCAmelCase_ ) ) def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] ): """simple docstring""" a :str = get_tokens(UpperCAmelCase_ ) a :List[Any] = get_tokens(UpperCAmelCase_ ) a :Dict = collections.Counter(UpperCAmelCase_ ) & collections.Counter(UpperCAmelCase_ ) a :Any = sum(common.values() ) if len(UpperCAmelCase_ ) == 0 or len(UpperCAmelCase_ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 a :int = 1.0 * num_same / len(UpperCAmelCase_ ) a :Tuple = 1.0 * num_same / len(UpperCAmelCase_ ) a :Any = (2 * precision * recall) / (precision + recall) return fa def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" a :Tuple = {} a :Any = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a :Tuple = qa['''id'''] a :int = [t for t in qa['''answers''']['''text'''] if normalize_answer(UpperCAmelCase_ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string a :List[Any] = [''''''] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue a :str = preds[qid] # Take max over all gold answers a :Optional[int] = max(compute_exact(UpperCAmelCase_ , UpperCAmelCase_ ) for a in gold_answers ) a :List[str] = max(compute_fa(UpperCAmelCase_ , UpperCAmelCase_ ) for a in gold_answers ) return exact_scores, fa_scores def __lowerCamelCase ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int ): """simple docstring""" a :List[Any] = {} for qid, s in scores.items(): a :Optional[int] = na_probs[qid] > na_prob_thresh if pred_na: a :List[str] = float(not qid_to_has_ans[qid] ) else: a :List[str] = s return new_scores def __lowerCamelCase ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Tuple=None ): """simple docstring""" if not qid_list: a :List[Any] = len(UpperCAmelCase_ ) return collections.OrderedDict( [ ('''exact''', 100.0 * sum(exact_scores.values() ) / total), ('''f1''', 100.0 * sum(fa_scores.values() ) / total), ('''total''', total), ] ) else: a :Tuple = len(UpperCAmelCase_ ) return collections.OrderedDict( [ ('''exact''', 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ('''f1''', 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ('''total''', total), ] ) def __lowerCamelCase ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : str ): """simple docstring""" for k in new_eval: a :Any = new_eval[k] def __lowerCamelCase ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] ): """simple docstring""" plt.step(UpperCAmelCase_ , UpperCAmelCase_ , color='''b''' , alpha=0.2 , where='''post''' ) plt.fill_between(UpperCAmelCase_ , UpperCAmelCase_ , step='''post''' , alpha=0.2 , color='''b''' ) plt.xlabel('''Recall''' ) plt.ylabel('''Precision''' ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(UpperCAmelCase_ ) plt.savefig(UpperCAmelCase_ ) plt.clf() def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Union[str, Any]=None , UpperCAmelCase_ : int=None ): """simple docstring""" a :Any = sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : na_probs[k] ) a :Optional[int] = 0.0 a :Optional[Any] = 1.0 a :Union[str, Any] = 0.0 a :Optional[int] = [1.0] a :Optional[int] = [0.0] a :Union[str, Any] = 0.0 for i, qid in enumerate(UpperCAmelCase_ ): if qid_to_has_ans[qid]: true_pos += scores[qid] a :List[str] = true_pos / float(i + 1 ) a :Any = true_pos / float(UpperCAmelCase_ ) if i == len(UpperCAmelCase_ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(UpperCAmelCase_ ) recalls.append(UpperCAmelCase_ ) if out_image: plot_pr_curve(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return {"ap": 100.0 * avg_prec} def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Tuple ): """simple docstring""" if out_image_dir and not os.path.exists(UpperCAmelCase_ ): os.makedirs(UpperCAmelCase_ ) a :str = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return a :List[str] = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_exact.png''' ) , title='''Precision-Recall curve for Exact Match score''' , ) a :Any = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_f1.png''' ) , title='''Precision-Recall curve for F1 score''' , ) a :List[Any] = {k: float(UpperCAmelCase_ ) for k, v in qid_to_has_ans.items()} a :Optional[Any] = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_oracle.png''' ) , title='''Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)''' , ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_exact''' ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_f1''' ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_oracle''' ) def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Tuple ): """simple docstring""" if not qid_list: return a :Tuple = [na_probs[k] for k in qid_list] a :Dict = np.ones_like(UpperCAmelCase_ ) / float(len(UpperCAmelCase_ ) ) plt.hist(UpperCAmelCase_ , weights=UpperCAmelCase_ , bins=20 , range=(0.0, 1.0) ) plt.xlabel('''Model probability of no-answer''' ) plt.ylabel('''Proportion of dataset''' ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(UpperCAmelCase_ , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[int] ): """simple docstring""" a :int = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) a :List[str] = num_no_ans a :List[str] = cur_score a :Dict = 0.0 a :Union[str, Any] = sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : na_probs[k] ) for i, qid in enumerate(UpperCAmelCase_ ): if qid not in scores: continue if qid_to_has_ans[qid]: a :Tuple = scores[qid] else: if preds[qid]: a :List[Any] = -1 else: a :Union[str, Any] = 0 cur_score += diff if cur_score > best_score: a :Any = cur_score a :Dict = na_probs[qid] return 100.0 * best_score / len(UpperCAmelCase_ ), best_thresh def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" a , a :str = find_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) a , a :Optional[Any] = find_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) a :Dict = best_exact a :Optional[int] = exact_thresh a :List[Any] = best_fa a :Optional[int] = fa_thresh def __lowerCamelCase ( ): """simple docstring""" with open(OPTS.data_file ) as f: a :Dict = json.load(UpperCAmelCase_ ) a :Optional[Any] = dataset_json['''data'''] with open(OPTS.pred_file ) as f: a :List[str] = json.load(UpperCAmelCase_ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: a :Any = json.load(UpperCAmelCase_ ) else: a :Optional[Any] = {k: 0.0 for k in preds} a :Tuple = make_qid_to_has_ans(UpperCAmelCase_ ) # maps qid to True/False a :Dict = [k for k, v in qid_to_has_ans.items() if v] a :Optional[Any] = [k for k, v in qid_to_has_ans.items() if not v] a , a :Union[str, Any] = get_raw_scores(UpperCAmelCase_ , UpperCAmelCase_ ) a :Tuple = apply_no_ans_threshold(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.na_prob_thresh ) a :Optional[Any] = apply_no_ans_threshold(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.na_prob_thresh ) a :Any = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ ) if has_ans_qids: a :Optional[int] = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ , qid_list=UpperCAmelCase_ ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''HasAns''' ) if no_ans_qids: a :Dict = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ , qid_list=UpperCAmelCase_ ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''NoAns''' ) if OPTS.na_prob_file: find_all_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir ) histogram_na_prob(UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir , '''hasAns''' ) histogram_na_prob(UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir , '''noAns''' ) if OPTS.out_file: with open(OPTS.out_file , '''w''' ) as f: json.dump(UpperCAmelCase_ , UpperCAmelCase_ ) else: print(json.dumps(UpperCAmelCase_ , indent=2 ) ) if __name__ == "__main__": snake_case : Any = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
281
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def __lowerCamelCase ( UpperCAmelCase_ : dict ): """simple docstring""" return (data["data"], data["target"]) def __lowerCamelCase ( UpperCAmelCase_ : np.ndarray , UpperCAmelCase_ : np.ndarray ): """simple docstring""" a :Optional[Any] = XGBClassifier() classifier.fit(UpperCAmelCase_ , UpperCAmelCase_ ) return classifier def __lowerCamelCase ( ): """simple docstring""" a :List[Any] = load_iris() a , a :Any = data_handling(UpperCAmelCase_ ) a , a , a , a :Tuple = train_test_split( UpperCAmelCase_ , UpperCAmelCase_ , test_size=0.25 ) a :List[Any] = iris['''target_names'''] # Create an XGBoost Classifier from the training data a :Optional[int] = xgboost(UpperCAmelCase_ , UpperCAmelCase_ ) # Display the confusion matrix of the classifier with both training and test sets ConfusionMatrixDisplay.from_estimator( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , display_labels=UpperCAmelCase_ , cmap='''Blues''' , normalize='''true''' , ) plt.title('''Normalized Confusion Matrix - IRIS Dataset''' ) plt.show() if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
281
1
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : UNetaDModel lowerCamelCase : KarrasVeScheduler def __init__( self : str , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : KarrasVeScheduler ) -> int: super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Dict , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : int = 5_0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , **UpperCAmelCase__ : Dict , ) -> Union[Tuple, ImagePipelineOutput]: lowerCAmelCase = self.unet.config.sample_size lowerCAmelCase = (batch_size, 3, img_size, img_size) lowerCAmelCase = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) lowerCAmelCase = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(UpperCAmelCase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper lowerCAmelCase = self.scheduler.schedule[t] lowerCAmelCase = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat lowerCAmelCase , lowerCAmelCase = self.scheduler.add_noise_to_input(UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. lowerCAmelCase = (sigma_hat / 2) * model((sample_hat + 1) / 2 , sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev lowerCAmelCase = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. lowerCAmelCase = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 , sigma_prev / 2 ).sample lowerCAmelCase = self.scheduler.step_correct( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , step_output.prev_sample , step_output['derivative'] , ) lowerCAmelCase = step_output.prev_sample lowerCAmelCase = (sample / 2 + 0.5).clamp(0 , 1 ) lowerCAmelCase = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
4
from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract lowerCamelCase : List[str] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( lowercase : Tuple , lowercase : Optional[int] , lowercase : Optional[int] ): '''simple docstring''' return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def _SCREAMING_SNAKE_CASE ( lowercase : np.ndarray , lowercase : Optional[str] , lowercase : Optional[str] ): '''simple docstring''' lowerCamelCase_ = to_pil_image(lowercase ) lowerCamelCase_ , lowerCamelCase_ = pil_image.size lowerCamelCase_ = pytesseract.image_to_data(lowercase , lang=lowercase , output_type='dict' , config=lowercase ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = data['text'], data['left'], data['top'], data['width'], data['height'] # filter empty words and corresponding coordinates lowerCamelCase_ = [idx for idx, word in enumerate(lowercase ) if not word.strip()] lowerCamelCase_ = [word for idx, word in enumerate(lowercase ) if idx not in irrelevant_indices] lowerCamelCase_ = [coord for idx, coord in enumerate(lowercase ) if idx not in irrelevant_indices] lowerCamelCase_ = [coord for idx, coord in enumerate(lowercase ) if idx not in irrelevant_indices] lowerCamelCase_ = [coord for idx, coord in enumerate(lowercase ) if idx not in irrelevant_indices] lowerCamelCase_ = [coord for idx, coord in enumerate(lowercase ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format lowerCamelCase_ = [] for x, y, w, h in zip(lowercase , lowercase , lowercase , lowercase ): lowerCamelCase_ = [x, y, x + w, y + h] actual_boxes.append(lowercase ) # finally, normalize the bounding boxes lowerCamelCase_ = [] for box in actual_boxes: normalized_boxes.append(normalize_box(lowercase , lowercase , lowercase ) ) assert len(lowercase ) == len(lowercase ), "Not as many words as there are bounding boxes" return words, normalized_boxes class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = ['''pixel_values'''] def __init__( self : int , A_ : bool = True , A_ : Dict[str, int] = None , A_ : PILImageResampling = PILImageResampling.BILINEAR , A_ : bool = True , A_ : float = 1 / 255 , A_ : bool = True , A_ : Union[float, Iterable[float]] = None , A_ : Union[float, Iterable[float]] = None , A_ : bool = True , A_ : Optional[str] = None , A_ : Optional[str] = "" , **A_ : Optional[int] , ) -> None: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = size if size is not None else {'height': 224, 'width': 224} lowerCamelCase_ = get_size_dict(A_ ) lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = resample lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_value lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD lowerCamelCase_ = apply_ocr lowerCamelCase_ = ocr_lang lowerCamelCase_ = tesseract_config def a__ ( self : str , A_ : np.ndarray , A_ : Dict[str, int] , A_ : PILImageResampling = PILImageResampling.BILINEAR , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : str , ) -> np.ndarray: """simple docstring""" lowerCamelCase_ = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) lowerCamelCase_ = (size['height'], size['width']) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def a__ ( self : Any , A_ : np.ndarray , A_ : Union[int, float] , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : Optional[Any] , ) -> np.ndarray: """simple docstring""" return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : np.ndarray , A_ : Union[float, Iterable[float]] , A_ : Union[float, Iterable[float]] , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : int , ) -> np.ndarray: """simple docstring""" return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def a__ ( self : List[Any] , A_ : ImageInput , A_ : bool = None , A_ : Dict[str, int] = None , A_ : Dict=None , A_ : bool = None , A_ : float = None , A_ : bool = None , A_ : Union[float, Iterable[float]] = None , A_ : Union[float, Iterable[float]] = None , A_ : bool = None , A_ : Optional[str] = None , A_ : Optional[str] = None , A_ : Optional[Union[str, TensorType]] = None , A_ : ChannelDimension = ChannelDimension.FIRST , **A_ : Any , ) -> PIL.Image.Image: """simple docstring""" lowerCamelCase_ = do_resize if do_resize is not None else self.do_resize lowerCamelCase_ = size if size is not None else self.size lowerCamelCase_ = get_size_dict(A_ ) lowerCamelCase_ = resample if resample is not None else self.resample lowerCamelCase_ = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase_ = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase_ = image_mean if image_mean is not None else self.image_mean lowerCamelCase_ = image_std if image_std is not None else self.image_std lowerCamelCase_ = apply_ocr if apply_ocr is not None else self.apply_ocr lowerCamelCase_ = ocr_lang if ocr_lang is not None else self.ocr_lang lowerCamelCase_ = tesseract_config if tesseract_config is not None else self.tesseract_config lowerCamelCase_ = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('If do_normalize is True, image_mean and image_std must be specified.' ) # All transformations expect numpy arrays. lowerCamelCase_ = [to_numpy_array(A_ ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , 'pytesseract' ) lowerCamelCase_ = [] lowerCamelCase_ = [] for image in images: lowerCamelCase_ , lowerCamelCase_ = apply_tesseract(A_ , A_ , A_ ) words_batch.append(A_ ) boxes_batch.append(A_ ) if do_resize: lowerCamelCase_ = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: lowerCamelCase_ = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: lowerCamelCase_ = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] lowerCamelCase_ = [to_channel_dimension_format(A_ , A_ ) for image in images] lowerCamelCase_ = BatchFeature(data={'pixel_values': images} , tensor_type=A_ ) if apply_ocr: lowerCamelCase_ = words_batch lowerCamelCase_ = boxes_batch return data
204
0
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( "--original_config_file", default=None, type=str, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--scheduler_type", default="pndm", type=str, help="Type of scheduler to use. Should be one of ['pndm', 'lms', 'ddim', 'euler', 'euler-ancestral', 'dpm']", ) parser.add_argument( "--pipeline_type", default=None, type=str, help=( "The pipeline type. One of 'FrozenOpenCLIPEmbedder', 'FrozenCLIPEmbedder', 'PaintByExample'" ". If `None` pipeline will be automatically inferred." ), ) parser.add_argument( "--image_size", default=None, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--prediction_type", default=None, type=str, help=( "The prediction type that the model was trained on. Use 'epsilon' for Stable Diffusion v1.X and Stable" " Diffusion v2 Base. Use 'v_prediction' for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") parser.add_argument( "--stable_unclip", type=str, default=None, required=False, help="Set if this is a stable unCLIP model. One of 'txt2img' or 'img2img'.", ) parser.add_argument( "--stable_unclip_prior", type=str, default=None, required=False, help="Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.", ) parser.add_argument( "--clip_stats_path", type=str, help="Path to the clip stats file. Only required if the stable unclip model's config specifies `model.params.noise_aug_config.params.clip_stats_path`.", required=False, ) parser.add_argument( "--controlnet", action="store_true", default=None, help="Set flag if this is a controlnet checkpoint." ) parser.add_argument("--half", action="store_true", help="Save weights in half precision.") parser.add_argument( "--vae_path", type=str, default=None, required=False, help="Set to a path, hub id to an already converted vae to not convert it again.", ) UpperCAmelCase__ = parser.parse_args() UpperCAmelCase__ = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
356
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = torch.device("cpu") def _a ( ) -> Union[str, Any]: a = '''http://images.cocodataset.org/val2017/000000039769.jpg''' a = Image.open(requests.get(a , stream=a ).raw ) return im def _a ( a :Dict ) -> Tuple: if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703e00, 2.1107e00, -2.0811e00, 8.8685e-01, 2.4360e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636e-01, 2.3478e-01, -1.6963e00, -1.7381e00, -8.6337e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768e-01, -4.7429e-01, -1.0897e00, -1.0248e00, 3.5523e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330e-01, 2.4211e-01, -6.0185e-01, -8.2789e-01, -6.0446e-02] ) def _a ( a :int , a :Any , a :Union[str, Any] ) -> int: a = dct.pop(a ) a = val def _a ( a :Any ) -> Dict: a = [] for k in state_dict.keys(): a = k if ".pwconv" in k: a = k_new.replace('''.pwconv''' , '''.point_wise_conv''' ) if ".dwconv" in k: a = k_new.replace('''.dwconv''' , '''.depth_wise_conv''' ) if ".Proj." in k: a = k_new.replace('''.Proj.''' , '''.proj.''' ) if "patch_embed" in k_new: a = k_new.replace('''patch_embed''' , '''swiftformer.patch_embed.patch_embedding''' ) if "network" in k_new: a = k_new.split('''.''' ) if ls[2].isdigit(): a = '''swiftformer.encoder.network.''' + ls[1] + '''.blocks.''' + ls[2] + '''.''' + '''.'''.join(ls[3:] ) else: a = k_new.replace('''network''' , '''swiftformer.encoder.network''' ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def _a ( a :List[Any] , a :Tuple , a :List[str] ) -> Union[str, Any]: a = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size a = 1_000 a = '''huggingface/label-files''' a = '''imagenet-1k-id2label.json''' a = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) a = {int(a ): v for k, v in idalabel.items()} a = idalabel a = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": a = [3, 3, 6, 4] a = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": a = [3, 3, 9, 6] a = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": a = [4, 3, 10, 5] a = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": a = [4, 4, 12, 6] a = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith('''https''' ): a = torch.hub.load_state_dict_from_url(a , map_location='''cpu''' , check_hash=a ) else: a = torch.load(a , map_location='''cpu''' ) a = checkpoint a = create_rename_keys(a ) for rename_key_src, rename_key_dest in rename_keys: rename_key(a , a , a ) # load HuggingFace model a = SwiftFormerForImageClassification(a ).eval() hf_model.load_state_dict(a ) # prepare test inputs a = prepare_img() a = ViTImageProcessor.from_pretrained('''preprocessor_config''' ) a = processor(images=a , return_tensors='''pt''' ) # compare outputs from both models a = get_expected_output(a ) a = hf_model(inputs['''pixel_values'''] ).logits assert hf_logits.shape == torch.Size([1, 1_000] ) assert torch.allclose(hf_logits[0, 0:5] , a , atol=1e-3 ) Path(a ).mkdir(exist_ok=a ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(a ) if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swiftformer_name", default="swiftformer_xs", choices=["swiftformer_xs", "swiftformer_s", "swiftformer_l1", "swiftformer_l3"], type=str, help="Name of the SwiftFormer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="./converted_outputs/", type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument("--original_ckpt", default=None, type=str, help="Path to the original model checkpoint.") UpperCAmelCase__ = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
26
0
def a__ ( ): return [list(range(10_00 - i ,-10_00 - i ,-1 ) ) for i in range(10_00 )] a_ = generate_large_matrix() a_ = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def a__ ( _UpperCamelCase : list[list[int]] ): assert all(row == sorted(_UpperCamelCase ,reverse=_UpperCamelCase ) for row in grid ) assert all(list(_UpperCamelCase ) == sorted(_UpperCamelCase ,reverse=_UpperCamelCase ) for col in zip(*_UpperCamelCase ) ) def a__ ( _UpperCamelCase : list[int] ): __lowerCamelCase = 0 __lowerCamelCase = len(_UpperCamelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: __lowerCamelCase = (left + right) // 2 __lowerCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: __lowerCamelCase = mid + 1 else: __lowerCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(_UpperCamelCase ) def a__ ( _UpperCamelCase : list[list[int]] ): __lowerCamelCase = 0 __lowerCamelCase = len(grid[0] ) for i in range(len(_UpperCamelCase ) ): __lowerCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(_UpperCamelCase ) * len(grid[0] )) - total def a__ ( _UpperCamelCase : list[list[int]] ): return len([number for row in grid for number in row if number < 0] ) def a__ ( _UpperCamelCase : list[list[int]] ): __lowerCamelCase = 0 for row in grid: for i, number in enumerate(_UpperCamelCase ): if number < 0: total += len(_UpperCamelCase ) - i break return total def a__ ( ): from timeit import timeit print('''Running benchmarks''' ) __lowerCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): __lowerCamelCase = timeit(F"""{func}(grid=grid)""" ,setup=_UpperCamelCase ,number=5_00 ) print(F"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
330
import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params a_ = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["""memory_attention""", """encoder_attn"""], ["""attention""", """attn"""], ["""/""", """."""], [""".LayerNorm.gamma""", """_layer_norm.weight"""], [""".LayerNorm.beta""", """_layer_norm.bias"""], ["""r.layer_""", """r.layers."""], ["""output_proj""", """out_proj"""], ["""ffn.dense_1.""", """fc2."""], ["""ffn.dense.""", """fc1."""], ["""ffn_layer_norm""", """final_layer_norm"""], ["""kernel""", """weight"""], ["""encoder_layer_norm.""", """encoder.layer_norm."""], ["""decoder_layer_norm.""", """decoder.layer_norm."""], ["""embeddings.weights""", """shared.weight"""], ] def a__ ( _UpperCamelCase : int ): for pegasus_name, hf_name in PATTERNS: __lowerCamelCase = k.replace(_UpperCamelCase ,_UpperCamelCase ) return k def a__ ( _UpperCamelCase : dict ,_UpperCamelCase : dict ): __lowerCamelCase = DEFAULTS.copy() cfg_kwargs.update(_UpperCamelCase ) __lowerCamelCase = PegasusConfig(**_UpperCamelCase ) __lowerCamelCase = PegasusForConditionalGeneration(_UpperCamelCase ) __lowerCamelCase = torch_model.model.state_dict() __lowerCamelCase = {} for k, v in tf_weights.items(): __lowerCamelCase = rename_state_dict_key(_UpperCamelCase ) if new_k not in sd: raise ValueError(F"""could not find new key {new_k} in state dict. (converted from {k})""" ) if "dense" in k or "proj" in new_k: __lowerCamelCase = v.T __lowerCamelCase = torch.tensor(_UpperCamelCase ,dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, F"""{new_k}, {k}, {v.shape}, {sd[new_k].shape}""" # make sure embedding.padding_idx is respected __lowerCamelCase = torch.zeros_like(mapping['''shared.weight'''][cfg.pad_token_id + 1] ) __lowerCamelCase = mapping['''shared.weight'''] __lowerCamelCase = mapping['''shared.weight'''] __lowerCamelCase = {k: torch.zeros_like(_UpperCamelCase ) for k, v in sd.items() if k.endswith('''bias''' ) and k not in mapping} mapping.update(**_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = torch_model.model.load_state_dict(_UpperCamelCase ,strict=_UpperCamelCase ) __lowerCamelCase = [ k for k in missing if k not in ['''encoder.embed_positions.weight''', '''decoder.embed_positions.weight'''] ] assert unexpected_missing == [], F"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], F"""no matches found for the following tf keys {extra}""" return torch_model def a__ ( _UpperCamelCase : str="./ckpt/aeslc/model.ckpt-32000" ): __lowerCamelCase = tf.train.list_variables(_UpperCamelCase ) __lowerCamelCase = {} __lowerCamelCase = ['''Adafactor''', '''global_step'''] for name, shape in tqdm(_UpperCamelCase ,desc='''converting tf checkpoint to dict''' ): __lowerCamelCase = any(pat in name for pat in ignore_name ) if skip_key: continue __lowerCamelCase = tf.train.load_variable(_UpperCamelCase ,_UpperCamelCase ) __lowerCamelCase = array return tf_weights def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str ): # save tokenizer first __lowerCamelCase = Path(_UpperCamelCase ).parent.name __lowerCamelCase = task_specific_params[F"""summarization_{dataset}"""]['''max_position_embeddings'''] __lowerCamelCase = PegasusTokenizer.from_pretrained('''sshleifer/pegasus''' ,model_max_length=_UpperCamelCase ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(_UpperCamelCase ) # convert model __lowerCamelCase = get_tf_weights_as_numpy(_UpperCamelCase ) __lowerCamelCase = task_specific_params[F"""summarization_{dataset}"""] if dataset == "large": __lowerCamelCase = task_specific_params __lowerCamelCase = convert_pegasus(_UpperCamelCase ,_UpperCamelCase ) torch_model.save_pretrained(_UpperCamelCase ) __lowerCamelCase = torch_model.state_dict() sd.pop('''model.decoder.embed_positions.weight''' ) sd.pop('''model.encoder.embed_positions.weight''' ) torch.save(_UpperCamelCase ,Path(_UpperCamelCase ) / '''pytorch_model.bin''' ) if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") a_ = parser.parse_args() if args.save_dir is None: a_ = Path(args.tf_ckpt_path).parent.name a_ = os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
330
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCAmelCase : Tuple = logging.get_logger(__name__) _UpperCAmelCase : int = { """Salesforce/instruct-blip-flan-t5""": """https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json""", } class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = '''instructblip_vision_model''' def __init__( self , snake_case=1408 , snake_case=6144 , snake_case=39 , snake_case=16 , snake_case=224 , snake_case=14 , snake_case="gelu" , snake_case=1e-6 , snake_case=0.0 , snake_case=1e-1_0 , snake_case=True , **snake_case , ): super().__init__(**snake_case ) snake_case_ = hidden_size snake_case_ = intermediate_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = patch_size snake_case_ = image_size snake_case_ = initializer_range snake_case_ = attention_dropout snake_case_ = layer_norm_eps snake_case_ = hidden_act snake_case_ = qkv_bias @classmethod def a ( cls , snake_case , **snake_case ): cls._set_token_in_kwargs(snake_case ) snake_case_ , snake_case_ = cls.get_config_dict(snake_case , **snake_case ) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": snake_case_ = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(snake_case , **snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Tuple = '''instructblip_qformer''' def __init__( self , snake_case=3_0522 , snake_case=768 , snake_case=12 , snake_case=12 , snake_case=3072 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=0.02 , snake_case=1e-1_2 , snake_case=0 , snake_case="absolute" , snake_case=2 , snake_case=1408 , **snake_case , ): super().__init__(pad_token_id=snake_case , **snake_case ) snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = position_embedding_type snake_case_ = cross_attention_frequency snake_case_ = encoder_hidden_size @classmethod def a ( cls , snake_case , **snake_case ): cls._set_token_in_kwargs(snake_case ) snake_case_ , snake_case_ = cls.get_config_dict(snake_case , **snake_case ) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": snake_case_ = config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(snake_case , **snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Tuple = '''instructblip''' __SCREAMING_SNAKE_CASE : Optional[Any] = True def __init__( self , snake_case=None , snake_case=None , snake_case=None , snake_case=32 , **snake_case ): super().__init__(**snake_case ) if vision_config is None: snake_case_ = {} logger.info('vision_config is None. initializing the InstructBlipVisionConfig with default values.' ) if qformer_config is None: snake_case_ = {} logger.info('qformer_config is None. Initializing the InstructBlipQFormerConfig with default values.' ) if text_config is None: snake_case_ = {} logger.info('text_config is None. Initializing the text config with default values (`OPTConfig`).' ) snake_case_ = InstructBlipVisionConfig(**snake_case ) snake_case_ = InstructBlipQFormerConfig(**snake_case ) snake_case_ = text_config['model_type'] if 'model_type' in text_config else 'opt' snake_case_ = CONFIG_MAPPING[text_model_type](**snake_case ) snake_case_ = self.text_config.tie_word_embeddings snake_case_ = self.text_config.is_encoder_decoder snake_case_ = num_query_tokens snake_case_ = self.vision_config.hidden_size snake_case_ = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES snake_case_ = 1.0 snake_case_ = 0.02 @classmethod def a ( cls , snake_case , snake_case , snake_case , **snake_case , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **snake_case , ) def a ( self ): snake_case_ = copy.deepcopy(self.__dict__ ) snake_case_ = self.vision_config.to_dict() snake_case_ = self.qformer_config.to_dict() snake_case_ = self.text_config.to_dict() snake_case_ = self.__class__.model_type return output
200
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json""" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : List[Any] = '''roformer''' def __init__( self , snake_case=5_0000 , snake_case=None , snake_case=768 , snake_case=12 , snake_case=12 , snake_case=3072 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=1536 , snake_case=2 , snake_case=0.02 , snake_case=1e-1_2 , snake_case=0 , snake_case=False , snake_case=True , **snake_case , ): super().__init__(pad_token_id=snake_case , **snake_case ) snake_case_ = vocab_size snake_case_ = hidden_size if embedding_size is None else embedding_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = rotary_value snake_case_ = use_cache class lowercase ( lowercase_ ): @property def a ( self ): if self.task == "multiple-choice": snake_case_ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: snake_case_ = {0: 'batch', 1: 'sequence'} snake_case_ = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
200
1
'''simple docstring''' import functools def _lowerCamelCase ( lowercase : list[int] , lowercase : list[int] ) -> int: # Validation if not isinstance(lowercase , lowercase ) or not all(isinstance(lowercase , lowercase ) for day in days ): raise ValueError("The parameter days should be a list of integers" ) if len(lowercase ) != 3 or not all(isinstance(lowercase , lowercase ) for cost in costs ): raise ValueError("The parameter costs should be a list of three integers" ) if len(lowercase ) == 0: return 0 if min(lowercase ) <= 0: raise ValueError("All days elements should be greater than 0" ) if max(lowercase ) >= 366: raise ValueError("All days elements should be less than 366" ) _a = set(lowercase ) @functools.cache def dynamic_programming(lowercase : int ) -> int: if index > 365: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 30 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
63
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable lowerCAmelCase_ : int = {'configuration_gpt_neox': ['GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoXConfig']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[int] = ['GPTNeoXTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ 'GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTNeoXForCausalLM', 'GPTNeoXForQuestionAnswering', 'GPTNeoXForSequenceClassification', 'GPTNeoXForTokenClassification', 'GPTNeoXLayer', 'GPTNeoXModel', 'GPTNeoXPreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys lowerCAmelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
63
1
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification _UpperCAmelCase : Union[str, Any] = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co _UpperCAmelCase : Union[str, Any] = '''main''' # Default branch name _UpperCAmelCase : Optional[Any] = '''f2c752cfc5c0ab6f4bdec59acea69eefbee381c2''' # One particular commit (not the top of `main`) _UpperCAmelCase : Tuple = '''aaaaaaa''' # This commit does not exist, so we should 404. _UpperCAmelCase : Dict = '''d9e9f15bc825e4b2c9249e9578f884bbcb5e3684''' # Sha-1 of config.json on the top of `main`, for checking purposes _UpperCAmelCase : Dict = '''4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3''' @contextlib.contextmanager def __lowerCamelCase ( ): '''simple docstring''' print('Welcome!' ) yield print('Bye!' ) @contextlib.contextmanager def __lowerCamelCase ( ): '''simple docstring''' print('Bonjour!' ) yield print('Au revoir!' ) class lowercase ( unittest.TestCase ): def a ( self ): # If the spec is missing, importlib would not be able to import the module dynamically. assert transformers.__spec__ is not None assert importlib.util.find_spec('transformers' ) is not None class lowercase ( unittest.TestCase ): @unittest.mock.patch('sys.stdout' , new_callable=io.StringIO ) def a ( self , snake_case ): with ContextManagers([] ): print('Transformers are awesome!' ) # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , 'Transformers are awesome!\n' ) @unittest.mock.patch('sys.stdout' , new_callable=io.StringIO ) def a ( self , snake_case ): with ContextManagers([context_en()] ): print('Transformers are awesome!' ) # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , 'Welcome!\nTransformers are awesome!\nBye!\n' ) @unittest.mock.patch('sys.stdout' , new_callable=io.StringIO ) def a ( self , snake_case ): with ContextManagers([context_fr(), context_en()] ): print('Transformers are awesome!' ) # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , 'Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n' ) @require_torch def a ( self ): self.assertEqual(find_labels(__snake_case ) , ['labels'] ) self.assertEqual(find_labels(__snake_case ) , ['labels', 'next_sentence_label'] ) self.assertEqual(find_labels(__snake_case ) , ['start_positions', 'end_positions'] ) class lowercase ( lowerCamelCase_ ): pass self.assertEqual(find_labels(__snake_case ) , ['labels'] ) @require_tf def a ( self ): self.assertEqual(find_labels(__snake_case ) , ['labels'] ) self.assertEqual(find_labels(__snake_case ) , ['labels', 'next_sentence_label'] ) self.assertEqual(find_labels(__snake_case ) , ['start_positions', 'end_positions'] ) class lowercase ( lowerCamelCase_ ): pass self.assertEqual(find_labels(__snake_case ) , ['labels'] ) @require_flax def a ( self ): # Flax models don't have labels self.assertEqual(find_labels(__snake_case ) , [] ) self.assertEqual(find_labels(__snake_case ) , [] ) self.assertEqual(find_labels(__snake_case ) , [] ) class lowercase ( lowerCamelCase_ ): pass self.assertEqual(find_labels(__snake_case ) , [] )
371
from __future__ import annotations import os from typing import Any import requests _UpperCAmelCase : int = """https://api.github.com""" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _UpperCAmelCase : Dict = BASE_URL + """/user""" # https://github.com/settings/tokens _UpperCAmelCase : Optional[Any] = os.environ.get("""USER_TOKEN""", """""") def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' snake_case_ = { 'Authorization': F'''token {auth_token}''', 'Accept': 'application/vnd.github.v3+json', } return requests.get(UpperCamelCase__ , headers=UpperCamelCase__ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F'''{key}: {value}''') else: raise ValueError("""'USER_TOKEN' field cannot be empty.""")
200
0
"""simple docstring""" import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _SCREAMING_SNAKE_CASE: def __init__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=13 ,SCREAMING_SNAKE_CASE__=[30, 30] ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=5 ,SCREAMING_SNAKE_CASE__=4 ,SCREAMING_SNAKE_CASE__=37 ,SCREAMING_SNAKE_CASE__="gelu" ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=10 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=None ,SCREAMING_SNAKE_CASE__=8 ,SCREAMING_SNAKE_CASE__=10 ,) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = parent __SCREAMING_SNAKE_CASE :List[Any] = batch_size __SCREAMING_SNAKE_CASE :int = image_size __SCREAMING_SNAKE_CASE :int = patch_size __SCREAMING_SNAKE_CASE :int = num_channels __SCREAMING_SNAKE_CASE :Optional[Any] = is_training __SCREAMING_SNAKE_CASE :List[str] = use_labels __SCREAMING_SNAKE_CASE :List[Any] = hidden_size __SCREAMING_SNAKE_CASE :str = num_hidden_layers __SCREAMING_SNAKE_CASE :Tuple = num_attention_heads __SCREAMING_SNAKE_CASE :List[str] = intermediate_size __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_act __SCREAMING_SNAKE_CASE :int = hidden_dropout_prob __SCREAMING_SNAKE_CASE :Optional[Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :str = type_sequence_label_size __SCREAMING_SNAKE_CASE :Dict = initializer_range __SCREAMING_SNAKE_CASE :Tuple = num_labels __SCREAMING_SNAKE_CASE :Dict = scope __SCREAMING_SNAKE_CASE :Dict = n_targets __SCREAMING_SNAKE_CASE :Optional[Any] = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens __SCREAMING_SNAKE_CASE :str = (image_size[1] // patch_size) * (image_size[0] // patch_size) __SCREAMING_SNAKE_CASE :Dict = num_patches + 1 + self.num_detection_tokens def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :str = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) __SCREAMING_SNAKE_CASE :Union[str, Any] = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) __SCREAMING_SNAKE_CASE :Any = [] for i in range(self.batch_size ): __SCREAMING_SNAKE_CASE :Dict = {} __SCREAMING_SNAKE_CASE :Union[str, Any] = torch.randint( high=self.num_labels ,size=(self.n_targets,) ,device=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = torch.rand(self.n_targets ,4 ,device=SCREAMING_SNAKE_CASE__ ) labels.append(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = self.get_config() return config, pixel_values, labels def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" return YolosConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=SCREAMING_SNAKE_CASE__ ,initializer_range=self.initializer_range ,num_detection_tokens=self.num_detection_tokens ,num_labels=self.num_labels ,) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = YolosModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() __SCREAMING_SNAKE_CASE :Dict = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.expected_seq_len, self.hidden_size) ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :int = YolosForObjectDetection(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() __SCREAMING_SNAKE_CASE :Optional[int] = model(pixel_values=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape ,(self.batch_size, self.num_detection_tokens, 4) ) __SCREAMING_SNAKE_CASE :Optional[Any] = model(pixel_values=SCREAMING_SNAKE_CASE__ ,labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.loss.shape ,() ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape ,(self.batch_size, self.num_detection_tokens, 4) ) def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Optional[int] = config_and_inputs __SCREAMING_SNAKE_CASE :Optional[int] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE( A , A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Dict = (YolosModel, YolosForObjectDetection) if is_torch_available() else () SCREAMING_SNAKE_CASE_ : Dict = ( {'''feature-extraction''': YolosModel, '''object-detection''': YolosForObjectDetection} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Dict = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Tuple = False def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=False ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :str = super()._prepare_for_class(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,return_labels=SCREAMING_SNAKE_CASE__ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": __SCREAMING_SNAKE_CASE :Dict = [] for i in range(self.model_tester.batch_size ): __SCREAMING_SNAKE_CASE :Union[str, Any] = {} __SCREAMING_SNAKE_CASE :Optional[Any] = torch.ones( size=(self.model_tester.n_targets,) ,device=SCREAMING_SNAKE_CASE__ ,dtype=torch.long ) __SCREAMING_SNAKE_CASE :Optional[Any] = torch.ones( self.model_tester.n_targets ,4 ,device=SCREAMING_SNAKE_CASE__ ,dtype=torch.float ) labels.append(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[str] = labels return inputs_dict def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :int = YolosModelTester(self ) __SCREAMING_SNAKE_CASE :Optional[Any] = ConfigTester(self ,config_class=SCREAMING_SNAKE_CASE__ ,has_text_modality=SCREAMING_SNAKE_CASE__ ,hidden_size=37 ) def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def _UpperCamelCase ( self ) -> Dict: """simple docstring""" pass def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE :Optional[int] = model_class(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __SCREAMING_SNAKE_CASE :List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE__ ,nn.Linear ) ) def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE :Union[str, Any] = model_class(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE :Optional[Any] = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE :Union[str, Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :int = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE :int = True # in YOLOS, the seq_len is different __SCREAMING_SNAKE_CASE :Optional[int] = self.model_tester.expected_seq_len for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE :str = True __SCREAMING_SNAKE_CASE :int = False __SCREAMING_SNAKE_CASE :Optional[Any] = True __SCREAMING_SNAKE_CASE :int = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE :str = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) ) __SCREAMING_SNAKE_CASE :Tuple = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) ,self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __SCREAMING_SNAKE_CASE :List[str] = True __SCREAMING_SNAKE_CASE :List[Any] = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE :Any = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) ) __SCREAMING_SNAKE_CASE :Tuple = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) __SCREAMING_SNAKE_CASE :List[str] = len(SCREAMING_SNAKE_CASE__ ) # Check attention is always last and order is fine __SCREAMING_SNAKE_CASE :str = True __SCREAMING_SNAKE_CASE :Union[str, Any] = True __SCREAMING_SNAKE_CASE :Union[str, Any] = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE :List[str] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) ) __SCREAMING_SNAKE_CASE :Dict = 1 self.assertEqual(out_len + added_hidden_states ,len(SCREAMING_SNAKE_CASE__ ) ) __SCREAMING_SNAKE_CASE :Optional[int] = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" def check_hidden_states_output(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ): __SCREAMING_SNAKE_CASE :Any = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE :Optional[Any] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) ) __SCREAMING_SNAKE_CASE :Tuple = outputs.hidden_states __SCREAMING_SNAKE_CASE :List[str] = getattr( self.model_tester ,'''expected_num_hidden_layers''' ,self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) ,SCREAMING_SNAKE_CASE__ ) # YOLOS has a different seq_length __SCREAMING_SNAKE_CASE :Optional[Any] = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[seq_length, self.model_tester.hidden_size] ,) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE :int = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE :Any = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*SCREAMING_SNAKE_CASE__ ) @slow def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE :Optional[Any] = YolosModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( ) -> str: __SCREAMING_SNAKE_CASE :Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _SCREAMING_SNAKE_CASE( unittest.TestCase ): @cached_property def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" return AutoImageProcessor.from_pretrained('''hustvl/yolos-small''' ) if is_vision_available() else None @slow def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = YolosForObjectDetection.from_pretrained('''hustvl/yolos-small''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Any = self.default_image_processor __SCREAMING_SNAKE_CASE :List[str] = prepare_img() __SCREAMING_SNAKE_CASE :List[Any] = image_processor(images=SCREAMING_SNAKE_CASE__ ,return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE :Union[str, Any] = model(inputs.pixel_values ) # verify outputs __SCREAMING_SNAKE_CASE :Optional[Any] = torch.Size((1, 1_00, 92) ) self.assertEqual(outputs.logits.shape ,SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = torch.tensor( [[-2_4.0_2_4_8, -1_0.3_0_2_4, -1_4.8_2_9_0], [-4_2.0_3_9_2, -1_6.8_2_0_0, -2_7.4_3_3_4], [-2_7.2_7_4_3, -1_1.8_1_5_4, -1_8.7_1_4_8]] ,device=SCREAMING_SNAKE_CASE__ ,) __SCREAMING_SNAKE_CASE :int = torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] ,device=SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] ,SCREAMING_SNAKE_CASE__ ,atol=1E-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] ,SCREAMING_SNAKE_CASE__ ,atol=1E-4 ) ) # verify postprocessing __SCREAMING_SNAKE_CASE :Any = image_processor.post_process_object_detection( SCREAMING_SNAKE_CASE__ ,threshold=0.3 ,target_sizes=[image.size[::-1]] )[0] __SCREAMING_SNAKE_CASE :Any = torch.tensor([0.9_9_9_4, 0.9_7_9_0, 0.9_9_6_4, 0.9_9_7_2, 0.9_8_6_1] ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = [75, 75, 17, 63, 17] __SCREAMING_SNAKE_CASE :List[str] = torch.tensor([3_3_5.0_6_0_9, 7_9.3_8_4_8, 3_7_5.4_2_1_6, 1_8_7.2_4_9_5] ).to(SCREAMING_SNAKE_CASE__ ) self.assertEqual(len(results['''scores'''] ) ,5 ) self.assertTrue(torch.allclose(results['''scores'''] ,SCREAMING_SNAKE_CASE__ ,atol=1E-4 ) ) self.assertSequenceEqual(results['''labels'''].tolist() ,SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(results['''boxes'''][0, :] ,SCREAMING_SNAKE_CASE__ ) )
191
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "facebook/xmod-base": "https://huggingface.co/facebook/xmod-base/resolve/main/config.json", "facebook/xmod-large-prenorm": "https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json", "facebook/xmod-base-13-125k": "https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json", "facebook/xmod-base-30-125k": "https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json", "facebook/xmod-base-30-195k": "https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json", "facebook/xmod-base-60-125k": "https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json", "facebook/xmod-base-60-265k": "https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json", "facebook/xmod-base-75-125k": "https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json", "facebook/xmod-base-75-269k": "https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json", } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : Any = '''xmod''' def __init__( self ,SCREAMING_SNAKE_CASE__=3_05_22 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu" ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=5_12 ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-12 ,SCREAMING_SNAKE_CASE__=1 ,SCREAMING_SNAKE_CASE__=0 ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__="absolute" ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=None ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=("en_XX",) ,SCREAMING_SNAKE_CASE__=None ,**SCREAMING_SNAKE_CASE__ ,) -> Optional[Any]: """simple docstring""" super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ ,bos_token_id=SCREAMING_SNAKE_CASE__ ,eos_token_id=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = vocab_size __SCREAMING_SNAKE_CASE :List[Any] = hidden_size __SCREAMING_SNAKE_CASE :List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE :List[str] = num_attention_heads __SCREAMING_SNAKE_CASE :Optional[int] = hidden_act __SCREAMING_SNAKE_CASE :Tuple = intermediate_size __SCREAMING_SNAKE_CASE :Dict = hidden_dropout_prob __SCREAMING_SNAKE_CASE :str = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Optional[Any] = max_position_embeddings __SCREAMING_SNAKE_CASE :Optional[Any] = type_vocab_size __SCREAMING_SNAKE_CASE :str = initializer_range __SCREAMING_SNAKE_CASE :List[Any] = layer_norm_eps __SCREAMING_SNAKE_CASE :Optional[Any] = position_embedding_type __SCREAMING_SNAKE_CASE :Any = use_cache __SCREAMING_SNAKE_CASE :List[str] = classifier_dropout __SCREAMING_SNAKE_CASE :Any = pre_norm __SCREAMING_SNAKE_CASE :Dict = adapter_reduction_factor __SCREAMING_SNAKE_CASE :Dict = adapter_layer_norm __SCREAMING_SNAKE_CASE :Dict = adapter_reuse_layer_norm __SCREAMING_SNAKE_CASE :Tuple = ln_before_adapter __SCREAMING_SNAKE_CASE :Any = list(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = default_language class _SCREAMING_SNAKE_CASE( A ): @property def _UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": __SCREAMING_SNAKE_CASE :Tuple = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __SCREAMING_SNAKE_CASE :Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
191
1
import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class a__ ( __lowercase ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaises(_A ): __lowerCAmelCase = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaises(_A ): __lowerCAmelCase = pa.array(TypedSequence([1, 2, 3] , try_type=Value("bool" ) , type=Value("int64" ) ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence([1, 2, 3] , type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): __lowerCAmelCase = pa.array(TypedSequence(["foo", "bar"] , type=Value("int64" ) ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence([1, 2, 3] , try_type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence(["foo", "bar"] , try_type=Value("int64" ) ) ) self.assertEqual(arr.type , pa.string() ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): __lowerCAmelCase = pa.array(TypedSequence(["foo", "bar"] , type=ArrayaD((1, 3) , "int64" ) ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = pa.array(TypedSequence(["foo", "bar"] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" import PIL.Image __lowerCAmelCase = PIL.Image.fromarray(np.arange(1_0 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( "datasets.arrow_writer.cast_to_python_objects" , side_effect=_A ) as mock_cast_to_python_objects: __lowerCAmelCase = pa.array(TypedSequence([{"path": None, "bytes": b"image_bytes"}, pil_image] , type=Image() ) ) __lowerCAmelCase = mock_cast_to_python_objects.call_args_list[-1] self.assertIn("optimize_list_casting" , _A ) self.assertFalse(kwargs["optimize_list_casting"] ) def _a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[int] ): __lowerCAmelCase = pa.BufferReader(_lowerCamelCase ) if isinstance(_lowerCamelCase , pa.Buffer ) else pa.memory_map(_lowerCamelCase ) __lowerCAmelCase = pa.ipc.open_stream(_lowerCamelCase ) __lowerCAmelCase = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def _a ( SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[str] ): __lowerCAmelCase = pa.BufferOutputStream() __lowerCAmelCase = pa.schema(_lowerCamelCase ) if fields else None with ArrowWriter(stream=_lowerCamelCase , schema=_lowerCamelCase , writer_batch_size=_lowerCamelCase ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __lowerCAmelCase = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(_lowerCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def _a ( ): __lowerCAmelCase = pa.BufferOutputStream() __lowerCAmelCase = Features({"labels": ClassLabel(names=["neg", "pos"] )} ) with ArrowWriter(stream=_lowerCamelCase , features=_lowerCamelCase ) as writer: writer.write({"labels": 0} ) writer.write({"labels": 1} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata __lowerCAmelCase = pa.BufferReader(output.getvalue() ) __lowerCAmelCase = pa.ipc.open_stream(_lowerCamelCase ) __lowerCAmelCase = f.read_all() __lowerCAmelCase = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(_lowerCamelCase ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) def _a ( SCREAMING_SNAKE_CASE_ : List[Any] ): __lowerCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=_lowerCamelCase , writer_batch_size=_lowerCamelCase , hash_salt="split_name" , check_duplicates=_lowerCamelCase , ) as writer: with pytest.raises(_lowerCamelCase ): writer.write({"col_1": "foo", "col_2": 1} , key=[1, 2] ) __lowerCAmelCase = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 10] ) def _a ( SCREAMING_SNAKE_CASE_ : Dict ): __lowerCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=_lowerCamelCase , writer_batch_size=_lowerCamelCase , hash_salt="split_name" , check_duplicates=_lowerCamelCase , ) as writer: with pytest.raises(_lowerCamelCase ): writer.write({"col_1": "foo", "col_2": 1} , key=10 ) writer.write({"col_1": "bar", "col_2": 2} , key=10 ) __lowerCAmelCase = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 10] ) def _a ( SCREAMING_SNAKE_CASE_ : Optional[int] ): __lowerCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=_lowerCamelCase , writer_batch_size=_lowerCamelCase , hash_salt="split_name" , check_duplicates=_lowerCamelCase , ) as writer: writer.write({"col_1": "foo", "col_2": 1} , key=1 ) writer.write({"col_1": "bar", "col_2": 2} , key=2 ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def _a ( SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Dict ): __lowerCAmelCase = pa.BufferOutputStream() __lowerCAmelCase = pa.schema(_lowerCamelCase ) if fields else None with ArrowWriter(stream=_lowerCamelCase , schema=_lowerCamelCase , writer_batch_size=_lowerCamelCase ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) writer.write_batch({"col_1": [], "col_2": []} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __lowerCAmelCase = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(_lowerCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def _a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] ): __lowerCAmelCase = pa.BufferOutputStream() __lowerCAmelCase = pa.schema(_lowerCamelCase ) if fields else None with ArrowWriter(stream=_lowerCamelCase , schema=_lowerCamelCase , writer_batch_size=_lowerCamelCase ) as writer: writer.write_table(pa.Table.from_pydict({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __lowerCAmelCase = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(_lowerCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def _a ( SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : List[str] ): __lowerCAmelCase = pa.BufferOutputStream() __lowerCAmelCase = pa.schema(_lowerCamelCase ) if fields else None with ArrowWriter(stream=_lowerCamelCase , schema=_lowerCamelCase , writer_batch_size=_lowerCamelCase ) as writer: writer.write_row(pa.Table.from_pydict({"col_1": ["foo"], "col_2": [1]} ) ) writer.write_row(pa.Table.from_pydict({"col_1": ["bar"], "col_2": [2]} ) ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __lowerCAmelCase = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(_lowerCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def _a ( ): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCAmelCase = {"""col_1""": pa.string(), """col_2""": pa.intaa()} __lowerCAmelCase = os.path.join(_lowerCamelCase , "test.arrow" ) with ArrowWriter(path=_lowerCamelCase , schema=pa.schema(_lowerCamelCase ) ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(_lowerCamelCase , metadata=writer._schema.metadata ) _check_output(_lowerCamelCase , 1 ) def _a ( SCREAMING_SNAKE_CASE_ : Optional[Any] ): if pa.types.is_list(_lowerCamelCase ): return get_base_dtype(arr_type.value_type ) else: return arr_type def _a ( SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str ): if isinstance(lst[0] , _lowerCamelCase ): change_first_primitive_element_in_list(lst[0] , _lowerCamelCase ) else: __lowerCAmelCase = value @pytest.mark.parametrize("optimized_int_type, expected_dtype" , [(None, pa.intaa()), (Value("int32" ), pa.intaa())] ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def _a ( SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict ): __lowerCAmelCase = pa.array(TypedSequence(_lowerCamelCase , optimized_int_type=_lowerCamelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( "col, expected_dtype" , [ ("attention_mask", pa.inta()), ("special_tokens_mask", pa.inta()), ("token_type_ids", pa.inta()), ("input_ids", pa.intaa()), ("other", pa.intaa()), ] , ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def _a ( SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : List[Any] ): __lowerCAmelCase = pa.array(OptimizedTypedSequence(_lowerCamelCase , col=_lowerCamelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications __lowerCAmelCase = copy.deepcopy(_lowerCamelCase ) __lowerCAmelCase = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(_lowerCamelCase , _lowerCamelCase ) __lowerCAmelCase = pa.array(OptimizedTypedSequence(_lowerCamelCase , col=_lowerCamelCase ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize("raise_exception" , [False, True] ) def _a ( SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : int ): __lowerCAmelCase = str(tmp_path / "dataset-train.arrow" ) try: with ArrowWriter(path=_lowerCamelCase ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def _a ( SCREAMING_SNAKE_CASE_ : Union[str, Any] ): __lowerCAmelCase = """mock://dataset-train.arrow""" with ArrowWriter(path=_lowerCamelCase , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(_lowerCamelCase ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(_lowerCamelCase ) def _a ( ): __lowerCAmelCase = pa.BufferOutputStream() with ParquetWriter(stream=_lowerCamelCase ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) __lowerCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 __lowerCAmelCase = pa.BufferReader(output.getvalue() ) __lowerCAmelCase = pq.read_table(_lowerCamelCase ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize("embed_local_files" , [False, True] ) def _a ( SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[int] ): import PIL.Image __lowerCAmelCase = str(tmp_path / "test_image_rgb.jpg" ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(_lowerCamelCase , format="png" ) __lowerCAmelCase = pa.BufferOutputStream() with ParquetWriter( stream=_lowerCamelCase , features=Features({"image": Image()} ) , embed_local_files=_lowerCamelCase ) as writer: writer.write({"image": image_path} ) writer.finalize() __lowerCAmelCase = pa.BufferReader(output.getvalue() ) __lowerCAmelCase = pq.read_table(_lowerCamelCase ) __lowerCAmelCase = pa_table.to_pydict() if embed_local_files: assert isinstance(out["image"][0]["path"] , _lowerCamelCase ) with open(_lowerCamelCase , "rb" ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def _a ( ): __lowerCAmelCase = pa.schema([pa.field("col_1" , pa.string() , nullable=_lowerCamelCase )] ) __lowerCAmelCase = pa.BufferOutputStream() with ArrowWriter(stream=_lowerCamelCase ) as writer: writer._build_writer(inferred_schema=_lowerCamelCase ) assert writer._schema == pa.schema([pa.field("col_1" , pa.string() )] )
363
from sklearn.metrics import mean_squared_error import datasets UpperCamelCase__ = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ UpperCamelCase__ = """\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. """ UpperCamelCase__ = """ Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {\"raw_values\", \"uniform_average\"} or array-like of shape (n_outputs,), default=\"uniform_average\" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. \"raw_values\" : Returns a full set of errors in case of multioutput input. \"uniform_average\" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric(\"mse\") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {'mse': 0.6123724356957945} If you're using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric(\"mse\", \"multilist\") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {'mse': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput='raw_values') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mse': array([0.41666667, 1. ])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ "https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html" ] , ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("float" ) ), "references": datasets.Sequence(datasets.Value("float" ) ), } else: return { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } def __SCREAMING_SNAKE_CASE( self , _A , _A , _A=None , _A="uniform_average" , _A=True ): """simple docstring""" __lowerCAmelCase = mean_squared_error( _A , _A , sample_weight=_A , multioutput=_A , squared=_A ) return {"mse": mse}
102
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : str = logging.get_logger(__name__) UpperCAmelCase_ : Any = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Tuple = '''beit''' def __init__( self : Dict , SCREAMING_SNAKE_CASE__ : Any=8_1_9_2 , SCREAMING_SNAKE_CASE__ : Tuple=7_6_8 , SCREAMING_SNAKE_CASE__ : Any=1_2 , SCREAMING_SNAKE_CASE__ : str=1_2 , SCREAMING_SNAKE_CASE__ : Any=3_0_7_2 , SCREAMING_SNAKE_CASE__ : Optional[int]="gelu" , SCREAMING_SNAKE_CASE__ : str=0.0 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.0 , SCREAMING_SNAKE_CASE__ : List[str]=0.02 , SCREAMING_SNAKE_CASE__ : Optional[int]=1E-12 , SCREAMING_SNAKE_CASE__ : Any=2_2_4 , SCREAMING_SNAKE_CASE__ : List[Any]=1_6 , SCREAMING_SNAKE_CASE__ : Dict=3 , SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : List[Any]=False , SCREAMING_SNAKE_CASE__ : str=False , SCREAMING_SNAKE_CASE__ : str=False , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : int=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=[3, 5, 7, 1_1] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=[1, 2, 3, 6] , SCREAMING_SNAKE_CASE__ : List[str]=True , SCREAMING_SNAKE_CASE__ : Any=0.4 , SCREAMING_SNAKE_CASE__ : Any=2_5_6 , SCREAMING_SNAKE_CASE__ : Any=1 , SCREAMING_SNAKE_CASE__ : Optional[int]=False , SCREAMING_SNAKE_CASE__ : List[Any]=2_5_5 , **SCREAMING_SNAKE_CASE__ : Optional[Any] , ) -> str: super().__init__(**SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = vocab_size a_ : str = hidden_size a_ : int = num_hidden_layers a_ : Any = num_attention_heads a_ : Any = intermediate_size a_ : Optional[Any] = hidden_act a_ : str = hidden_dropout_prob a_ : List[Any] = attention_probs_dropout_prob a_ : Optional[Any] = initializer_range a_ : Any = layer_norm_eps a_ : Dict = image_size a_ : List[str] = patch_size a_ : Optional[int] = num_channels a_ : str = use_mask_token a_ : Optional[int] = use_absolute_position_embeddings a_ : Union[str, Any] = use_relative_position_bias a_ : Optional[Any] = use_shared_relative_position_bias a_ : Optional[int] = layer_scale_init_value a_ : str = drop_path_rate a_ : Tuple = use_mean_pooling # decode head attributes (semantic segmentation) a_ : str = out_indices a_ : List[str] = pool_scales # auxiliary head attributes (semantic segmentation) a_ : Optional[int] = use_auxiliary_head a_ : Optional[Any] = auxiliary_loss_weight a_ : List[str] = auxiliary_channels a_ : List[str] = auxiliary_num_convs a_ : Dict = auxiliary_concat_input a_ : str = semantic_loss_ignore_index class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Dict = version.parse('''1.11''' ) @property def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> float: return 1E-4
32
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ : Optional[int] = { 'facebook/mask2former-swin-small-coco-instance': ( 'https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json' ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } UpperCAmelCase_ : List[str] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Any = '''mask2former''' snake_case__ : Any = ['''swin'''] snake_case__ : str = {'''hidden_size''': '''hidden_dim'''} def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Optional[Dict] = None , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 1_0_2_4 , SCREAMING_SNAKE_CASE__ : str = "relu" , SCREAMING_SNAKE_CASE__ : int = 6 , SCREAMING_SNAKE_CASE__ : int = 1_0 , SCREAMING_SNAKE_CASE__ : int = 8 , SCREAMING_SNAKE_CASE__ : float = 0.0 , SCREAMING_SNAKE_CASE__ : int = 2_0_4_8 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : int = 4 , SCREAMING_SNAKE_CASE__ : int = 2_5_5 , SCREAMING_SNAKE_CASE__ : int = 1_0_0 , SCREAMING_SNAKE_CASE__ : float = 0.1 , SCREAMING_SNAKE_CASE__ : float = 2.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : int = 1_2_5_4_4 , SCREAMING_SNAKE_CASE__ : float = 3.0 , SCREAMING_SNAKE_CASE__ : float = 0.75 , SCREAMING_SNAKE_CASE__ : float = 0.02 , SCREAMING_SNAKE_CASE__ : float = 1.0 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : List[int] = [4, 8, 1_6, 3_2] , SCREAMING_SNAKE_CASE__ : bool = None , **SCREAMING_SNAKE_CASE__ : int , ) -> List[Any]: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) a_ : Dict = CONFIG_MAPPING['swin']( image_size=2_2_4 , in_channels=3 , patch_size=4 , embed_dim=9_6 , depths=[2, 2, 1_8, 2] , num_heads=[3, 6, 1_2, 2_4] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=SCREAMING_SNAKE_CASE__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): a_ : Any = backbone_config.pop('model_type' ) a_ : Optional[Any] = CONFIG_MAPPING[backbone_model_type] a_ : List[str] = config_class.from_dict(SCREAMING_SNAKE_CASE__ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. """ F"""Supported model types: {",".join(self.backbones_supported )}""" ) a_ : Dict = backbone_config a_ : List[str] = feature_size a_ : List[str] = mask_feature_size a_ : int = hidden_dim a_ : Dict = encoder_feedforward_dim a_ : str = activation_function a_ : List[str] = encoder_layers a_ : List[str] = decoder_layers a_ : Dict = num_attention_heads a_ : str = dropout a_ : Tuple = dim_feedforward a_ : List[str] = pre_norm a_ : Optional[int] = enforce_input_projection a_ : Any = common_stride a_ : Optional[int] = ignore_value a_ : int = num_queries a_ : Tuple = no_object_weight a_ : Dict = class_weight a_ : Optional[int] = mask_weight a_ : Optional[int] = dice_weight a_ : str = train_num_points a_ : List[str] = oversample_ratio a_ : List[Any] = importance_sample_ratio a_ : Any = init_std a_ : Union[str, Any] = init_xavier_std a_ : Union[str, Any] = use_auxiliary_loss a_ : Dict = feature_strides a_ : List[str] = output_auxiliary_logits a_ : Dict = decoder_layers super().__init__(**SCREAMING_SNAKE_CASE__ ) @classmethod def SCREAMING_SNAKE_CASE ( cls : str , SCREAMING_SNAKE_CASE__ : PretrainedConfig , **SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> List[Any]: return cls( backbone_config=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict[str, any]: a_ : Optional[int] = copy.deepcopy(self.__dict__ ) a_ : List[Any] = self.backbone_config.to_dict() a_ : Optional[Any] = self.__class__.model_type return output
32
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer _lowercase : str = logging.get_logger(__name__) _lowercase : Optional[int] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} # See all MVP models at https://huggingface.co/models?filter=mvp _lowercase : Optional[int] = { "vocab_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json", }, "added_tokens.json": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json", }, "merges_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt", }, "tokenizer_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json", }, } _lowercase : str = { "RUCAIBox/mvp": 1_0_2_4, } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = ['input_ids', 'attention_mask'] _a = MvpTokenizer def __init__( self : Optional[Any], lowerCamelCase : Any=None, lowerCamelCase : List[str]=None, lowerCamelCase : int=None, lowerCamelCase : int="replace", lowerCamelCase : List[Any]="<s>", lowerCamelCase : Optional[Any]="</s>", lowerCamelCase : Dict="</s>", lowerCamelCase : Union[str, Any]="<s>", lowerCamelCase : int="<unk>", lowerCamelCase : List[str]="<pad>", lowerCamelCase : List[str]="<mask>", lowerCamelCase : Union[str, Any]=False, lowerCamelCase : Dict=True, **lowerCamelCase : Optional[Any], )-> Optional[Any]: super().__init__( lowerCamelCase, lowerCamelCase, tokenizer_file=lowerCamelCase, errors=lowerCamelCase, bos_token=lowerCamelCase, eos_token=lowerCamelCase, sep_token=lowerCamelCase, cls_token=lowerCamelCase, unk_token=lowerCamelCase, pad_token=lowerCamelCase, mask_token=lowerCamelCase, add_prefix_space=lowerCamelCase, trim_offsets=lowerCamelCase, **lowerCamelCase, ) lowerCamelCase__ : List[Any] =json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''', lowerCamelCase ) != add_prefix_space: lowerCamelCase__ : Optional[Any] =getattr(lowerCamelCase, pre_tok_state.pop('''type''' ) ) lowerCamelCase__ : str =add_prefix_space lowerCamelCase__ : List[Any] =pre_tok_class(**lowerCamelCase ) lowerCamelCase__ : str =add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCamelCase__ : int ='''post_processor''' lowerCamelCase__ : Optional[Any] =getattr(self.backend_tokenizer, lowerCamelCase, lowerCamelCase ) if tokenizer_component_instance: lowerCamelCase__ : int =json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : Optional[int] =tuple(state['''sep'''] ) if "cls" in state: lowerCamelCase__ : str =tuple(state['''cls'''] ) lowerCamelCase__ : List[Any] =False if state.get('''add_prefix_space''', lowerCamelCase ) != add_prefix_space: lowerCamelCase__ : Optional[Any] =add_prefix_space lowerCamelCase__ : Dict =True if state.get('''trim_offsets''', lowerCamelCase ) != trim_offsets: lowerCamelCase__ : List[str] =trim_offsets lowerCamelCase__ : List[Any] =True if changes_to_apply: lowerCamelCase__ : Tuple =getattr(lowerCamelCase, state.pop('''type''' ) ) lowerCamelCase__ : Optional[Any] =component_class(**lowerCamelCase ) setattr(self.backend_tokenizer, lowerCamelCase, lowerCamelCase ) @property def snake_case ( self : Tuple )-> str: if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def snake_case ( self : Tuple, lowerCamelCase : List[Any] )-> List[str]: lowerCamelCase__ : Dict =AddedToken(lowerCamelCase, lstrip=lowerCamelCase, rstrip=lowerCamelCase ) if isinstance(lowerCamelCase, lowerCamelCase ) else value lowerCamelCase__ : Union[str, Any] =value def snake_case ( self : Tuple, *lowerCamelCase : Optional[Any], **lowerCamelCase : int )-> BatchEncoding: lowerCamelCase__ : Any =kwargs.get('''is_split_into_words''', lowerCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*lowerCamelCase, **lowerCamelCase ) def snake_case ( self : Any, *lowerCamelCase : str, **lowerCamelCase : Dict )-> BatchEncoding: lowerCamelCase__ : Dict =kwargs.get('''is_split_into_words''', lowerCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*lowerCamelCase, **lowerCamelCase ) def snake_case ( self : List[str], lowerCamelCase : str, lowerCamelCase : Optional[str] = None )-> Tuple[str]: lowerCamelCase__ : str =self._tokenizer.model.save(lowerCamelCase, name=lowerCamelCase ) return tuple(lowerCamelCase ) def snake_case ( self : Optional[int], lowerCamelCase : List[Any], lowerCamelCase : List[str]=None )-> int: lowerCamelCase__ : Optional[Any] =[self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def snake_case ( self : List[Any], lowerCamelCase : List[int], lowerCamelCase : Optional[List[int]] = None )-> List[int]: lowerCamelCase__ : Optional[Any] =[self.sep_token_id] lowerCamelCase__ : int =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
370
"""simple docstring""" from collections import defaultdict class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Union[str, Any], lowerCamelCase : List[Any], lowerCamelCase : List[str] )-> Optional[int]: lowerCamelCase__ : List[Any] =total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 lowerCamelCase__ : Optional[Any] =[ [-1 for i in range(total + 1 )] for j in range(2 ** len(lowerCamelCase ) ) ] lowerCamelCase__ : Any =defaultdict(lowerCamelCase ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 lowerCamelCase__ : List[Any] =(1 << len(lowerCamelCase )) - 1 def snake_case ( self : int, lowerCamelCase : str, lowerCamelCase : Any )-> Any: # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement lowerCamelCase__ : Optional[int] =self.count_ways_until(lowerCamelCase, task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p), task_no + 1 ) # save the value. lowerCamelCase__ : int =total_ways_util return self.dp[mask][task_no] def snake_case ( self : Dict, lowerCamelCase : Dict )-> int: # Store the list of persons for each task for i in range(len(lowerCamelCase ) ): for j in task_performed[i]: self.task[j].append(lowerCamelCase ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0, 1 ) if __name__ == "__main__": _lowercase : Tuple = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. _lowercase : Dict = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
272
0
print((lambda quine: quine % quine)("print((lambda quine: quine %% quine)(%r))"))
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase__ = { "configuration_nllb_moe": [ "NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP", "NllbMoeConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST", "NllbMoeForConditionalGeneration", "NllbMoeModel", "NllbMoePreTrainedModel", "NllbMoeTop2Router", "NllbMoeSparseMLP", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
0
1
"""simple docstring""" import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> List[Any]: '''simple docstring''' super().__init__() UpperCAmelCase : Tuple = value_function UpperCAmelCase : Dict = unet UpperCAmelCase : Union[str, Any] = scheduler UpperCAmelCase : List[Any] = env UpperCAmelCase : int = env.get_dataset() UpperCAmelCase : Optional[int] = {} for key in self.data.keys(): try: UpperCAmelCase : Dict = self.data[key].mean() except: # noqa: E722 pass UpperCAmelCase : int = {} for key in self.data.keys(): try: UpperCAmelCase : Optional[Any] = self.data[key].std() except: # noqa: E722 pass UpperCAmelCase : Any = env.observation_space.shape[0] UpperCAmelCase : str = env.action_space.shape[0] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' return (x_in - self.means[key]) / self.stds[key] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' return x_in * self.stds[key] + self.means[key] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' if type(_SCREAMING_SNAKE_CASE ) is dict: return {k: self.to_torch(_SCREAMING_SNAKE_CASE ) for k, v in x_in.items()} elif torch.is_tensor(_SCREAMING_SNAKE_CASE ): return x_in.to(self.unet.device ) return torch.tensor(_SCREAMING_SNAKE_CASE , device=self.unet.device ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' for key, val in cond.items(): UpperCAmelCase : Optional[Any] = val.clone() return x_in def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' UpperCAmelCase : Dict = x.shape[0] UpperCAmelCase : Optional[int] = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model UpperCAmelCase : Tuple = torch.full((batch_size,) , _SCREAMING_SNAKE_CASE , device=self.unet.device , dtype=torch.long ) for _ in range(_SCREAMING_SNAKE_CASE ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models UpperCAmelCase : Dict = self.value_function(x.permute(0 , 2 , 1 ) , _SCREAMING_SNAKE_CASE ).sample UpperCAmelCase : Optional[int] = torch.autograd.grad([y.sum()] , [x] )[0] UpperCAmelCase : List[Any] = self.scheduler._get_variance(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[str] = torch.exp(0.5 * posterior_variance ) UpperCAmelCase : str = model_std * grad UpperCAmelCase : str = 0 UpperCAmelCase : Any = x.detach() UpperCAmelCase : int = x + scale * grad UpperCAmelCase : Any = self.reset_xa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.action_dim ) UpperCAmelCase : Optional[int] = self.unet(x.permute(0 , 2 , 1 ) , _SCREAMING_SNAKE_CASE ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg UpperCAmelCase : Any = self.scheduler.step(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , predict_epsilon=_SCREAMING_SNAKE_CASE )["""prev_sample"""] # apply conditions to the trajectory (set the initial state) UpperCAmelCase : Dict = self.reset_xa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.action_dim ) UpperCAmelCase : int = self.to_torch(_SCREAMING_SNAKE_CASE ) return x, y def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=64 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.1 ) -> Tuple: '''simple docstring''' UpperCAmelCase : int = self.normalize(_SCREAMING_SNAKE_CASE , """observations""" ) UpperCAmelCase : int = obs[None].repeat(_SCREAMING_SNAKE_CASE , axis=0 ) UpperCAmelCase : Dict = {0: self.to_torch(_SCREAMING_SNAKE_CASE )} UpperCAmelCase : Dict = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) UpperCAmelCase : str = randn_tensor(_SCREAMING_SNAKE_CASE , device=self.unet.device ) UpperCAmelCase : Any = self.reset_xa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.action_dim ) UpperCAmelCase : str = self.to_torch(_SCREAMING_SNAKE_CASE ) # run the diffusion process UpperCAmelCase : Any = self.run_diffusion(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # sort output trajectories by value UpperCAmelCase : List[str] = y.argsort(0 , descending=_SCREAMING_SNAKE_CASE ).squeeze() UpperCAmelCase : Any = x[sorted_idx] UpperCAmelCase : Dict = sorted_values[:, :, : self.action_dim] UpperCAmelCase : int = actions.detach().cpu().numpy() UpperCAmelCase : List[str] = self.de_normalize(_SCREAMING_SNAKE_CASE , key="""actions""" ) # select the action with the highest value if y is not None: UpperCAmelCase : Any = 0 else: # if we didn't run value guiding, select a random action UpperCAmelCase : Optional[int] = np.random.randint(0 , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Any = denorm_actions[selected_index, 0] return denorm_actions
366
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer A: int = logging.get_logger(__name__) A: Any = {"vocab_file": "vocab.txt"} A: Optional[int] = { "vocab_file": { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt", } } A: Optional[int] = { "YituTech/conv-bert-base": 5_1_2, "YituTech/conv-bert-medium-small": 5_1_2, "YituTech/conv-bert-small": 5_1_2, } A: int = { "YituTech/conv-bert-base": {"do_lower_case": True}, "YituTech/conv-bert-medium-small": {"do_lower_case": True}, "YituTech/conv-bert-small": {"do_lower_case": True}, } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : List[Any] = VOCAB_FILES_NAMES __lowerCAmelCase : Any = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase : List[str] = PRETRAINED_INIT_CONFIGURATION __lowerCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase : int = ConvBertTokenizer def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE="[UNK]" , _SCREAMING_SNAKE_CASE="[SEP]" , _SCREAMING_SNAKE_CASE="[PAD]" , _SCREAMING_SNAKE_CASE="[CLS]" , _SCREAMING_SNAKE_CASE="[MASK]" , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE , ) -> Union[str, Any]: '''simple docstring''' super().__init__( _SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , tokenize_chinese_chars=_SCREAMING_SNAKE_CASE , strip_accents=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) UpperCAmelCase : int = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _SCREAMING_SNAKE_CASE ) != do_lower_case or normalizer_state.get("""strip_accents""" , _SCREAMING_SNAKE_CASE ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _SCREAMING_SNAKE_CASE ) != tokenize_chinese_chars ): UpperCAmelCase : Dict = getattr(_SCREAMING_SNAKE_CASE , normalizer_state.pop("""type""" ) ) UpperCAmelCase : str = do_lower_case UpperCAmelCase : Optional[int] = strip_accents UpperCAmelCase : List[str] = tokenize_chinese_chars UpperCAmelCase : Dict = normalizer_class(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase : int = do_lower_case def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) -> Tuple: '''simple docstring''' UpperCAmelCase : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[int]: '''simple docstring''' UpperCAmelCase : str = [self.sep_token_id] UpperCAmelCase : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: '''simple docstring''' UpperCAmelCase : Dict = self._tokenizer.model.save(_SCREAMING_SNAKE_CASE , name=_SCREAMING_SNAKE_CASE ) return tuple(_SCREAMING_SNAKE_CASE )
76
0
import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : List[str] ) -> int: assert isinstance(a_ , a_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : str , snake_case__ : int ) -> Tuple: UpperCamelCase : str = tmp_path / """cache""" UpperCamelCase : int = {"""text""": """string"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : Optional[Any] = TextDatasetReader(a_ , cache_dir=a_ , keep_in_memory=a_ ).read() _check_text_dataset(a_ , a_ ) @pytest.mark.parametrize( 'features' , [ None, {'text': 'string'}, {'text': 'int32'}, {'text': 'float32'}, ] , ) def UpperCamelCase ( snake_case__ : Any , snake_case__ : Tuple , snake_case__ : Union[str, Any] ) -> Optional[int]: UpperCamelCase : Any = tmp_path / """cache""" UpperCamelCase : Dict = {"""text""": """string"""} UpperCamelCase : Any = features.copy() if features else default_expected_features UpperCamelCase : Union[str, Any] = ( Features({feature: Value(a_ ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : List[str] = TextDatasetReader(a_ , features=a_ , cache_dir=a_ ).read() _check_text_dataset(a_ , a_ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : Tuple ) -> Union[str, Any]: UpperCamelCase : List[Any] = tmp_path / """cache""" UpperCamelCase : Optional[int] = {"""text""": """string"""} UpperCamelCase : List[str] = TextDatasetReader(a_ , cache_dir=a_ , split=a_ ).read() _check_text_dataset(a_ , a_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('path_type' , [str, list] ) def UpperCamelCase ( snake_case__ : str , snake_case__ : Any , snake_case__ : str ) -> Tuple: if issubclass(a_ , a_ ): UpperCamelCase : Optional[Any] = text_path elif issubclass(a_ , a_ ): UpperCamelCase : Tuple = [text_path] UpperCamelCase : str = tmp_path / """cache""" UpperCamelCase : Optional[Any] = {"""text""": """string"""} UpperCamelCase : Union[str, Any] = TextDatasetReader(a_ , cache_dir=a_ ).read() _check_text_dataset(a_ , a_ ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Optional[Any] , snake_case__ : Optional[Any]=("train",) ) -> Optional[int]: assert isinstance(a_ , a_ ) for split in splits: UpperCamelCase : Optional[int] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def UpperCamelCase ( snake_case__ : List[Any] , snake_case__ : Dict , snake_case__ : Dict ) -> Optional[Any]: UpperCamelCase : Optional[int] = tmp_path / """cache""" UpperCamelCase : List[str] = {"""text""": """string"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : int = TextDatasetReader({'train': text_path} , cache_dir=a_ , keep_in_memory=a_ ).read() _check_text_datasetdict(a_ , a_ ) @pytest.mark.parametrize( 'features' , [ None, {'text': 'string'}, {'text': 'int32'}, {'text': 'float32'}, ] , ) def UpperCamelCase ( snake_case__ : int , snake_case__ : List[str] , snake_case__ : Optional[int] ) -> Union[str, Any]: UpperCamelCase : Union[str, Any] = tmp_path / """cache""" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" UpperCamelCase : Tuple = {"""text""": """string"""} UpperCamelCase : Optional[Any] = features.copy() if features else default_expected_features UpperCamelCase : Dict = ( Features({feature: Value(a_ ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : List[str] = TextDatasetReader({'train': text_path} , features=a_ , cache_dir=a_ ).read() _check_text_datasetdict(a_ , a_ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Any , snake_case__ : Tuple ) -> Optional[Any]: if split: UpperCamelCase : int = {split: text_path} else: UpperCamelCase : str = """train""" UpperCamelCase : Any = {"""train""": text_path, """test""": text_path} UpperCamelCase : List[Any] = tmp_path / """cache""" UpperCamelCase : Dict = {"""text""": """string"""} UpperCamelCase : Optional[Any] = TextDatasetReader(a_ , cache_dir=a_ ).read() _check_text_datasetdict(a_ , a_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
'''simple docstring''' import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard @pytest.mark.parametrize( 'compression_format, is_archive' , [ ('7z', True), ('bz2', False), ('gzip', False), ('lz4', False), ('tar', True), ('xz', False), ('zip', True), ('zstd', False), ] , ) def lowerCamelCase ( lowerCAmelCase : List[Any] , lowerCAmelCase : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : Dict , lowerCAmelCase : Any , lowerCAmelCase : Optional[Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Tuple , lowerCAmelCase : Any , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Any , ): """simple docstring""" __magic_name__ : List[str] = { '7z': (seven_zip_file, SevenZipExtractor), 'bz2': (bza_file, BzipaExtractor), 'gzip': (gz_file, GzipExtractor), 'lz4': (lza_file, LzaExtractor), 'tar': (tar_file, TarExtractor), 'xz': (xz_file, XzExtractor), 'zip': (zip_file, ZipExtractor), 'zstd': (zstd_file, ZstdExtractor), } __magic_name__ , __magic_name__ : Optional[Any] = input_paths_and_base_extractors[compression_format] if input_path is None: __magic_name__ : Tuple = f'for \'{compression_format}\' compression_format, ' if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(UpperCAmelCase__ ) assert base_extractor.is_extractable(UpperCAmelCase__ ) __magic_name__ : int = tmp_path / ('extracted' if is_archive else 'extracted.txt') base_extractor.extract(UpperCAmelCase__ , UpperCAmelCase__ ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __magic_name__ : Any = file_path.read_text(encoding='utf-8' ) else: __magic_name__ : Union[str, Any] = output_path.read_text(encoding='utf-8' ) __magic_name__ : Optional[Any] = text_file.read_text(encoding='utf-8' ) assert extracted_file_content == expected_file_content @pytest.mark.parametrize( 'compression_format, is_archive' , [ ('7z', True), ('bz2', False), ('gzip', False), ('lz4', False), ('tar', True), ('xz', False), ('zip', True), ('zstd', False), ] , ) def lowerCamelCase ( lowerCAmelCase : int , lowerCAmelCase : Any , lowerCAmelCase : Optional[Any] , lowerCAmelCase : int , lowerCAmelCase : int , lowerCAmelCase : Optional[int] , lowerCAmelCase : int , lowerCAmelCase : str , lowerCAmelCase : List[Any] , lowerCAmelCase : Any , lowerCAmelCase : Tuple , lowerCAmelCase : Tuple , ): """simple docstring""" __magic_name__ : Optional[int] = { '7z': seven_zip_file, 'bz2': bza_file, 'gzip': gz_file, 'lz4': lza_file, 'tar': tar_file, 'xz': xz_file, 'zip': zip_file, 'zstd': zstd_file, } __magic_name__ : int = input_paths[compression_format] if input_path is None: __magic_name__ : Tuple = f'for \'{compression_format}\' compression_format, ' if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(UpperCAmelCase__ ) __magic_name__ : str = Extractor.infer_extractor_format(UpperCAmelCase__ ) assert extractor_format is not None __magic_name__ : str = tmp_path / ('extracted' if is_archive else 'extracted.txt') Extractor.extract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __magic_name__ : Dict = file_path.read_text(encoding='utf-8' ) else: __magic_name__ : List[str] = output_path.read_text(encoding='utf-8' ) __magic_name__ : List[str] = text_file.read_text(encoding='utf-8' ) assert extracted_file_content == expected_file_content @pytest.fixture def lowerCamelCase ( lowerCAmelCase : str , lowerCAmelCase : List[str] ): """simple docstring""" import tarfile __magic_name__ : List[Any] = tmp_path / 'data_dot_dot' directory.mkdir() __magic_name__ : int = directory / 'tar_file_with_dot_dot.tar' with tarfile.TarFile(UpperCAmelCase__ , 'w' ) as f: f.add(UpperCAmelCase__ , arcname=os.path.join('..' , text_file.name ) ) return path @pytest.fixture def lowerCamelCase ( lowerCAmelCase : Optional[int] ): """simple docstring""" import tarfile __magic_name__ : Union[str, Any] = tmp_path / 'data_sym_link' directory.mkdir() __magic_name__ : Tuple = directory / 'tar_file_with_sym_link.tar' os.symlink('..' , directory / 'subdir' , target_is_directory=UpperCAmelCase__ ) with tarfile.TarFile(UpperCAmelCase__ , 'w' ) as f: f.add(str(directory / 'subdir' ) , arcname='subdir' ) # str required by os.readlink on Windows and Python < 3.8 return path @pytest.mark.parametrize( 'insecure_tar_file, error_log' , [('tar_file_with_dot_dot', 'illegal path'), ('tar_file_with_sym_link', 'Symlink')] , ) def lowerCamelCase ( lowerCAmelCase : List[Any] , lowerCAmelCase : str , lowerCAmelCase : int , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : int , lowerCAmelCase : Optional[int] ): """simple docstring""" __magic_name__ : List[str] = { 'tar_file_with_dot_dot': tar_file_with_dot_dot, 'tar_file_with_sym_link': tar_file_with_sym_link, } __magic_name__ : Optional[int] = insecure_tar_files[insecure_tar_file] __magic_name__ : Tuple = tmp_path / 'extracted' TarExtractor.extract(UpperCAmelCase__ , UpperCAmelCase__ ) assert caplog.text for record in caplog.records: assert record.levelname == "ERROR" assert error_log in record.msg def lowerCamelCase ( lowerCAmelCase : List[Any] ): """simple docstring""" __magic_name__ : str = tmpdir / 'not_a_zip_file' # From: https://github.com/python/cpython/pull/5053 __magic_name__ : Any = ( b'\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x01\x00\x00' b'\x00\x02\x08\x06\x00\x00\x00\x99\x81\xb6\'\x00\x00\x00\x15I' b'DATx\x01\x01\n\x00\xf5\xff\x00PK\x05\x06\x00PK\x06\x06\x07' b'\xac\x01N\xc6|a\r\x00\x00\x00\x00IEND\xaeB`\x82' ) with not_a_zip_file.open('wb' ) as f: f.write(UpperCAmelCase__ ) assert zipfile.is_zipfile(str(UpperCAmelCase__ ) ) # is a false positive for `zipfile` assert not ZipExtractor.is_extractable(UpperCAmelCase__ ) # but we're right
363
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCamelCase ( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' A_ : int = StableDiffusionXLImgaImgPipeline A_ : str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} A_ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} A_ : Any = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ : Any = IMAGE_TO_IMAGE_IMAGE_PARAMS A_ : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __lowerCAmelCase ( self : Dict ) -> Optional[int]: torch.manual_seed(0 ) __magic_name__ : int = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , attention_head_dim=(2, 4) , use_linear_projection=_A , addition_embed_type='text_time' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __magic_name__ : str = EulerDiscreteScheduler( beta_start=0.0_0085 , beta_end=0.012 , steps_offset=1 , beta_schedule='scaled_linear' , timestep_spacing='leading' , ) torch.manual_seed(0 ) __magic_name__ : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __magic_name__ : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=32 , ) __magic_name__ : Dict = CLIPTextModel(_A ) __magic_name__ : List[str] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=_A ) __magic_name__ : Optional[Any] = CLIPTextModelWithProjection(_A ) __magic_name__ : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=_A ) __magic_name__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def __lowerCAmelCase ( self : List[Any] , _A : List[str] , _A : Any=0 ) -> Union[str, Any]: __magic_name__ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) __magic_name__ : Dict = image / 2 + 0.5 if str(_A ).startswith('mps' ): __magic_name__ : Any = torch.manual_seed(_A ) else: __magic_name__ : int = torch.Generator(device=_A ).manual_seed(_A ) __magic_name__ : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.75, } return inputs def __lowerCAmelCase ( self : Optional[Any] ) -> List[Any]: __magic_name__ : str = 'cpu' # ensure determinism for the device-dependent torch.Generator __magic_name__ : str = self.get_dummy_components() __magic_name__ : Any = StableDiffusionXLImgaImgPipeline(**_A ) __magic_name__ : List[Any] = sd_pipe.to(_A ) sd_pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Any = self.get_dummy_inputs(_A ) __magic_name__ : Optional[int] = sd_pipe(**_A ).images __magic_name__ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __magic_name__ : Any = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCAmelCase ( self : Optional[int] ) -> Optional[Any]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def __lowerCAmelCase ( self : List[Any] ) -> int: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def __lowerCAmelCase ( self : Any ) -> Union[str, Any]: pass def __lowerCAmelCase ( self : List[Any] ) -> Optional[int]: __magic_name__ : Dict = self.get_dummy_components() __magic_name__ : Optional[Any] = StableDiffusionXLImgaImgPipeline(**_A ) __magic_name__ : List[Any] = sd_pipe.to(_A ) __magic_name__ : str = sd_pipe.to(_A ) sd_pipe.set_progress_bar_config(disable=_A ) # forward without prompt embeds __magic_name__ : Union[str, Any] = self.get_dummy_inputs(_A ) __magic_name__ : Union[str, Any] = 3 * ['this is a negative prompt'] __magic_name__ : List[str] = negative_prompt __magic_name__ : int = 3 * [inputs['prompt']] __magic_name__ : Tuple = sd_pipe(**_A ) __magic_name__ : str = output.images[0, -3:, -3:, -1] # forward with prompt embeds __magic_name__ : Optional[Any] = self.get_dummy_inputs(_A ) __magic_name__ : Tuple = 3 * ['this is a negative prompt'] __magic_name__ : List[str] = 3 * [inputs.pop('prompt' )] ( ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ) : List[Any] = sd_pipe.encode_prompt(_A , negative_prompt=_A ) __magic_name__ : Tuple = sd_pipe( **_A , prompt_embeds=_A , negative_prompt_embeds=_A , pooled_prompt_embeds=_A , negative_pooled_prompt_embeds=_A , ) __magic_name__ : int = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : List[Any] ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : str , _A : Optional[int] , _A : Optional[Any]="cpu" , _A : List[str]=torch.floataa , _A : Any=0 ) -> str: __magic_name__ : List[str] = torch.Generator(device=_A ).manual_seed(_A ) __magic_name__ : Optional[Any] = np.random.RandomState(_A ).standard_normal((1, 4, 64, 64) ) __magic_name__ : Union[str, Any] = torch.from_numpy(_A ).to(device=_A , dtype=_A ) __magic_name__ : Optional[int] = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict: __magic_name__ : str = DiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base' ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Optional[int] = self.get_inputs(_A ) __magic_name__ : Union[str, Any] = pipe(**_A ).images __magic_name__ : Any = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __magic_name__ : List[Any] = np.array([0.4_9493, 0.4_7896, 0.4_0798, 0.5_4214, 0.5_3212, 0.4_8202, 0.4_7656, 0.4_6329, 0.4_8506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
275
0
'''simple docstring''' from functools import reduce __a: List[Any] = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __UpperCamelCase ( UpperCAmelCase = N ): return max( # mypy cannot properly interpret reduce int(reduce(lambda UpperCAmelCase , UpperCAmelCase : str(int(__UpperCamelCase ) * int(__UpperCamelCase ) ) , n[i : i + 13] ) ) for i in range(len(__UpperCamelCase ) - 12 ) ) if __name__ == "__main__": print(F'{solution() = }')
198
import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) A : Optional[int] = logging.getLogger(__name__) A : int = "Hello world! cécé herlolip" A : List[Any] = namedtuple( "BertAbsConfig", [ "temp_dir", "large", "use_bert_emb", "finetune_bert", "encoder", "share_emb", "max_pos", "enc_layers", "enc_hidden_size", "enc_heads", "enc_ff_size", "enc_dropout", "dec_layers", "dec_hidden_size", "dec_heads", "dec_ff_size", "dec_dropout", ], ) def a__ ( __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = BertAbsConfig( temp_dir="." , finetune_bert=__UpperCamelCase , large=__UpperCamelCase , share_emb=__UpperCamelCase , use_bert_emb=__UpperCamelCase , encoder="bert" , max_pos=5_1_2 , enc_layers=6 , enc_hidden_size=5_1_2 , enc_heads=8 , enc_ff_size=5_1_2 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=7_6_8 , dec_heads=8 , dec_ff_size=2_0_4_8 , dec_dropout=0.2 , ) SCREAMING_SNAKE_CASE_ = torch.load(__UpperCamelCase , lambda __UpperCamelCase , __UpperCamelCase : storage ) SCREAMING_SNAKE_CASE_ = AbsSummarizer(__UpperCamelCase , torch.device("cpu" ) , __UpperCamelCase ) original.eval() SCREAMING_SNAKE_CASE_ = BertAbsSummarizer(__UpperCamelCase , torch.device("cpu" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("convert the model" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("Make sure that the models' outputs are identical" ) SCREAMING_SNAKE_CASE_ = BertTokenizer.from_pretrained("bert-base-uncased" ) # prepare the model inputs SCREAMING_SNAKE_CASE_ = tokenizer.encode("This is sample éàalj'-." ) encoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(__UpperCamelCase )) ) SCREAMING_SNAKE_CASE_ = torch.tensor(__UpperCamelCase ).unsqueeze(0 ) SCREAMING_SNAKE_CASE_ = tokenizer.encode("This is sample 3 éàalj'-." ) decoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(__UpperCamelCase )) ) SCREAMING_SNAKE_CASE_ = torch.tensor(__UpperCamelCase ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass SCREAMING_SNAKE_CASE_ = encoder_input_ids SCREAMING_SNAKE_CASE_ = decoder_input_ids SCREAMING_SNAKE_CASE_ = SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical SCREAMING_SNAKE_CASE_ = original(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )[0] SCREAMING_SNAKE_CASE_ = original.generator(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = new_model( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )[0] SCREAMING_SNAKE_CASE_ = new_model.generator(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(__UpperCamelCase ) ) SCREAMING_SNAKE_CASE_ = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(__UpperCamelCase ) ) SCREAMING_SNAKE_CASE_ = torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-3 ) if are_identical: logging.info("all weights are equal up to 1e-3" ) else: raise ValueError("the weights are different. The new model is likely different from the original one." ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("saving the model's state dictionary" ) torch.save( new_model.state_dict() , "./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin" ) if __name__ == "__main__": A : Any = argparse.ArgumentParser() parser.add_argument( "--bertabs_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model.", ) A : int = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
118
0
'''simple docstring''' def UpperCamelCase ( _lowerCamelCase : int ): if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): raise TypeError("Input value must be a 'int' type" ) return bin(_lowerCamelCase ).count("1" ) if __name__ == "__main__": import doctest doctest.testmod()
123
'''simple docstring''' def UpperCamelCase ( _lowerCamelCase : int ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): A__ = F"Input value of [number={number}] must be an integer" raise TypeError(_lowerCamelCase ) if number < 1: A__ = F"Input value of [number={number}] must be > 0" raise ValueError(_lowerCamelCase ) A__ = 1 for i in range(1 , _lowerCamelCase ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
123
1
from math import factorial def _lowerCAmelCase ( lowerCAmelCase_ :int = 20 )->int: '''simple docstring''' snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(lowerCAmelCase_ ) / (factorial(lowerCAmelCase_ ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: SCREAMING_SNAKE_CASE :List[str] = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
159
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __lowerCAmelCase : """simple docstring""" def __init__( self : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int=1_3 , _lowerCAmelCase : Optional[int]=3 , _lowerCAmelCase : Any=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Any=2_2_4 , _lowerCAmelCase : Any=1_0_0_0 , _lowerCAmelCase : Any=[3, 3, 6, 4] , _lowerCAmelCase : Any=[4_8, 5_6, 1_1_2, 2_2_0] , ) -> List[Any]: """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = is_training snake_case_ = use_labels snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = num_labels snake_case_ = image_size snake_case_ = layer_depths snake_case_ = embed_dims def lowerCAmelCase__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=_lowerCAmelCase , layer_scale_init_value=1e-5 , ) def lowerCAmelCase__ ( self : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> str: """simple docstring""" snake_case_ = SwiftFormerModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def lowerCAmelCase__ ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> List[Any]: """simple docstring""" snake_case_ = self.num_labels snake_case_ = SwiftFormerForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) snake_case_ = SwiftFormerForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self : int ) -> Any: """simple docstring""" ((snake_case_) , (snake_case_) , (snake_case_)) = self.prepare_config_and_inputs() snake_case_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( a , a , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () _SCREAMING_SNAKE_CASE = ( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False def lowerCAmelCase__ ( self : Optional[int] ) -> Any: """simple docstring""" snake_case_ = SwiftFormerModelTester(self ) snake_case_ = ConfigTester( self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=3_7 , num_attention_heads=1_2 , num_hidden_layers=1_2 , ) def lowerCAmelCase__ ( self : Any ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def lowerCAmelCase__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" pass def lowerCAmelCase__ ( self : List[str] ) -> List[str]: """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(_lowerCAmelCase ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def lowerCAmelCase__ ( self : List[str] ) -> Any: """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(_lowerCAmelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def lowerCAmelCase__ ( self : str ) -> Optional[Any]: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self : str ) -> Any: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def lowerCAmelCase__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwiftFormerModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def lowerCAmelCase__ ( self : Any ) -> Tuple: """simple docstring""" pass def lowerCAmelCase__ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple ): snake_case_ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 8 self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(_lowerCAmelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase__ ( self : Any ) -> Optional[int]: """simple docstring""" def _config_zero_init(_lowerCAmelCase : List[str] ): snake_case_ = copy.deepcopy(_lowerCAmelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(_lowerCAmelCase , _lowerCAmelCase , 1e-10 ) if isinstance(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase ): snake_case_ = _config_zero_init(getattr(_lowerCAmelCase , _lowerCAmelCase ) ) setattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return configs_no_init snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(_lowerCAmelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=_lowerCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def lowerCAmelCase__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def _lowerCAmelCase ( )->str: '''simple docstring''' snake_case_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def lowerCAmelCase__ ( self : int ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def lowerCAmelCase__ ( self : List[str] ) -> Optional[int]: """simple docstring""" snake_case_ = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(_lowerCAmelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=_lowerCAmelCase , return_tensors="pt" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**_lowerCAmelCase ) # verify the logits snake_case_ = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) snake_case_ = torch.tensor([[-2.1703e00, 2.1107e00, -2.0811e00]] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) )
159
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_informer': [ 'INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'InformerForPrediction', 'InformerModel', 'InformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
361
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase_ = len(__UpperCamelCase ) while cur > 1: # Find the maximum number in arr UpperCAmelCase_ = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi UpperCAmelCase_ = arr[mi::-1] + arr[mi + 1 : len(__UpperCamelCase )] # Reverse whole list UpperCAmelCase_ = arr[cur - 1 :: -1] + arr[cur : len(__UpperCamelCase )] cur -= 1 return arr if __name__ == "__main__": _lowerCamelCase = input('Enter numbers separated by a comma:\n').strip() _lowerCamelCase = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
177
0
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' while b: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = b, a % b return a def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(lowerCAmelCase_ , a % b ) def UpperCAmelCase__ (): '''simple docstring''' print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
54
"""simple docstring""" from __future__ import annotations import pandas as pd def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [0] * no_of_processes __SCREAMING_SNAKE_CASE = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = burst_time[i] __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 9_9999_9999 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = False # Process until all processes are completed while complete != no_of_processes: for j in range(lowerCAmelCase_ ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: __SCREAMING_SNAKE_CASE = remaining_time[j] __SCREAMING_SNAKE_CASE = j __SCREAMING_SNAKE_CASE = True if not check: increment_time += 1 continue remaining_time[short] -= 1 __SCREAMING_SNAKE_CASE = remaining_time[short] if minm == 0: __SCREAMING_SNAKE_CASE = 9_9999_9999 if remaining_time[short] == 0: complete += 1 __SCREAMING_SNAKE_CASE = False # Find finish time of current process __SCREAMING_SNAKE_CASE = increment_time + 1 # Calculate waiting time __SCREAMING_SNAKE_CASE = finish_time - arrival_time[short] __SCREAMING_SNAKE_CASE = finar - burst_time[short] if waiting_time[short] < 0: __SCREAMING_SNAKE_CASE = 0 # Increment time increment_time += 1 return waiting_time def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [0] * no_of_processes for i in range(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = burst_time[i] + waiting_time[i] return turn_around_time def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 for i in range(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = total_waiting_time + waiting_time[i] __SCREAMING_SNAKE_CASE = total_turn_around_time + turn_around_time[i] print(f"""Average waiting time = {total_waiting_time / no_of_processes:.5f}""" ) print("Average turn around time =" , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print('''Enter how many process you want to analyze''') a__ : Optional[Any] = int(input()) a__ : Optional[int] = [0] * no_of_processes a__ : int = [0] * no_of_processes a__ : List[Any] = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print('''Enter the arrival time and burst time for process:--''' + str(i + 1)) a__ , a__ : Tuple = map(int, input().split()) a__ : int = calculate_waitingtime(arrival_time, burst_time, no_of_processes) a__ : Dict = burst_time a__ : Any = no_of_processes a__ : Optional[int] = waiting_time a__ : Union[str, Any] = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) a__ : str = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ '''Process''', '''BurstTime''', '''ArrivalTime''', '''WaitingTime''', '''TurnAroundTime''', ], ) # Printing the dataFrame pd.set_option('''display.max_rows''', fcfs.shape[0] + 1) print(fcfs)
54
1
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCamelCase__ ( A : Dict , A : str ): '''simple docstring''' UpperCAmelCase = checkpoint UpperCAmelCase = {} UpperCAmelCase = vae_state_dict['''encoder.conv_in.weight'''] UpperCAmelCase = vae_state_dict['''encoder.conv_in.bias'''] UpperCAmelCase = vae_state_dict['''encoder.conv_out.weight'''] UpperCAmelCase = vae_state_dict['''encoder.conv_out.bias'''] UpperCAmelCase = vae_state_dict['''encoder.norm_out.weight'''] UpperCAmelCase = vae_state_dict['''encoder.norm_out.bias'''] UpperCAmelCase = vae_state_dict['''decoder.conv_in.weight'''] UpperCAmelCase = vae_state_dict['''decoder.conv_in.bias'''] UpperCAmelCase = vae_state_dict['''decoder.conv_out.weight'''] UpperCAmelCase = vae_state_dict['''decoder.conv_out.bias'''] UpperCAmelCase = vae_state_dict['''decoder.norm_out.weight'''] UpperCAmelCase = vae_state_dict['''decoder.norm_out.bias'''] UpperCAmelCase = vae_state_dict['''quant_conv.weight'''] UpperCAmelCase = vae_state_dict['''quant_conv.bias'''] UpperCAmelCase = vae_state_dict['''post_quant_conv.weight'''] UpperCAmelCase = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only UpperCAmelCase = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) UpperCAmelCase = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(A ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) UpperCAmelCase = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(A ) } for i in range(A ): UpperCAmelCase = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase = renew_vae_resnet_paths(A ) UpperCAmelCase = {'''old''': f"""down.{i}.block""", '''new''': f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) UpperCAmelCase = [key for key in vae_state_dict if '''encoder.mid.block''' in key] UpperCAmelCase = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase = renew_vae_resnet_paths(A ) UpperCAmelCase = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) UpperCAmelCase = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] UpperCAmelCase = renew_vae_attention_paths(A ) UpperCAmelCase = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) conv_attn_to_linear(A ) for i in range(A ): UpperCAmelCase = num_up_blocks - 1 - i UpperCAmelCase = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase = renew_vae_resnet_paths(A ) UpperCAmelCase = {'''old''': f"""up.{block_id}.block""", '''new''': f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) UpperCAmelCase = [key for key in vae_state_dict if '''decoder.mid.block''' in key] UpperCAmelCase = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase = renew_vae_resnet_paths(A ) UpperCAmelCase = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) UpperCAmelCase = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] UpperCAmelCase = renew_vae_attention_paths(A ) UpperCAmelCase = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(A , A , A , additional_replacements=[meta_path] , config=A ) conv_attn_to_linear(A ) return new_checkpoint def lowerCamelCase__ ( A : str , A : str , ): '''simple docstring''' UpperCAmelCase = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) UpperCAmelCase = io.BytesIO(r.content ) UpperCAmelCase = OmegaConf.load(A ) UpperCAmelCase = 5_12 UpperCAmelCase = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open UpperCAmelCase = {} with safe_open(A , framework='''pt''' , device='''cpu''' ) as f: for key in f.keys(): UpperCAmelCase = f.get_tensor(A ) else: UpperCAmelCase = torch.load(A , map_location=A )['''state_dict'''] # Convert the VAE model. UpperCAmelCase = create_vae_diffusers_config(A , image_size=A ) UpperCAmelCase = custom_convert_ldm_vae_checkpoint(A , A ) UpperCAmelCase = AutoencoderKL(**A ) vae.load_state_dict(A ) vae.save_pretrained(A ) if __name__ == "__main__": _lowercase : Tuple = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") _lowercase : Dict = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
352
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class UpperCamelCase__: __magic_name__ : List[str] __magic_name__ : Optional[str] = None # Automatically constructed __magic_name__ : ClassVar[str] = "dict" __magic_name__ : ClassVar[Any] = None __magic_name__ : str = field(default="Translation" , init=lowerCAmelCase , repr=lowerCAmelCase ) def __call__( self : Union[str, Any] )-> str: """simple docstring""" return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def a__( self : int )-> Union["FeatureType", Dict[str, "FeatureType"]]: """simple docstring""" from .features import Value return {k: Value('''string''' ) for k in sorted(self.languages )} @dataclass class UpperCamelCase__: __magic_name__ : Optional[List] = None __magic_name__ : Optional[int] = None __magic_name__ : Optional[str] = None # Automatically constructed __magic_name__ : ClassVar[str] = "dict" __magic_name__ : ClassVar[Any] = None __magic_name__ : str = field(default="TranslationVariableLanguages" , init=lowerCAmelCase , repr=lowerCAmelCase ) def a__( self : Union[str, Any] )-> Optional[Any]: """simple docstring""" UpperCAmelCase = sorted(set(self.languages ) ) if self.languages else None UpperCAmelCase = len(self.languages ) if self.languages else None def __call__( self : int )-> Optional[Any]: """simple docstring""" return pa.struct({'''language''': pa.list_(pa.string() ), '''translation''': pa.list_(pa.string() )} ) def a__( self : Optional[int] , lowerCAmelCase : Dict )-> Tuple: """simple docstring""" UpperCAmelCase = set(self.languages ) if self.languages and set(lowerCAmelCase ) - lang_set: raise ValueError( F"""Some languages in example ({", ".join(sorted(set(lowerCAmelCase ) - lang_set ) )}) are not in valid set ({", ".join(lowerCAmelCase )}).""" ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. UpperCAmelCase = [] for lang, text in translation_dict.items(): if isinstance(lowerCAmelCase , lowerCAmelCase ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. UpperCAmelCase , UpperCAmelCase = zip(*sorted(lowerCAmelCase ) ) return {"language": languages, "translation": translations} def a__( self : Any )-> Union["FeatureType", Dict[str, "FeatureType"]]: """simple docstring""" from .features import Sequence, Value return { "language": Sequence(Value('''string''' ) ), "translation": Sequence(Value('''string''' ) ), }
91
0
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase__ (UpperCamelCase_ = 3 ): """simple docstring""" if isinstance(UpperCamelCase_ ,UpperCamelCase_ ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(UpperCamelCase_ ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) snake_case = QuantumRegister(UpperCamelCase_ ,'''qr''' ) snake_case = ClassicalRegister(UpperCamelCase_ ,'''cr''' ) snake_case = QuantumCircuit(UpperCamelCase_ ,UpperCamelCase_ ) snake_case = number_of_qubits for i in range(UpperCamelCase_ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(UpperCamelCase_ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) ,UpperCamelCase_ ,UpperCamelCase_ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(UpperCamelCase_ ,number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(UpperCamelCase_ ,UpperCamelCase_ ) # simulate with 10000 shots snake_case = Aer.get_backend('''qasm_simulator''' ) snake_case = execute(UpperCamelCase_ ,UpperCamelCase_ ,shots=1_00_00 ) return job.result().get_counts(UpperCamelCase_ ) if __name__ == "__main__": print( f'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
127
import argparse import math import traceback import dateutil.parser as date_parser import requests def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" snake_case = {} snake_case = job['''started_at'''] snake_case = job['''completed_at'''] snake_case = date_parser.parse(UpperCamelCase_ ) snake_case = date_parser.parse(UpperCamelCase_ ) snake_case = round((end_datetime - start_datetime).total_seconds() / 60.0 ) snake_case = start snake_case = end snake_case = duration_in_min return job_info def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_=None ): """simple docstring""" snake_case = None if token is not None: snake_case = {'''Accept''': '''application/vnd.github+json''', '''Authorization''': F'''Bearer {token}'''} snake_case = F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' snake_case = requests.get(UpperCamelCase_ ,headers=UpperCamelCase_ ).json() snake_case = {} try: job_time.update({job['''name''']: extract_time_from_single_job(UpperCamelCase_ ) for job in result['''jobs''']} ) snake_case = math.ceil((result['''total_count'''] - 1_00) / 1_00 ) for i in range(UpperCamelCase_ ): snake_case = requests.get(url + F'''&page={i + 2}''' ,headers=UpperCamelCase_ ).json() job_time.update({job['''name''']: extract_time_from_single_job(UpperCamelCase_ ) for job in result['''jobs''']} ) return job_time except Exception: print(F'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") _SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() _SCREAMING_SNAKE_CASE : int = get_job_time(args.workflow_run_id) _SCREAMING_SNAKE_CASE : Dict = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f'''{k}: {v['duration']}''')
127
1
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _lowerCamelCase = logging.getLogger(__name__) def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : str ) -> List[str]: """simple docstring""" return (preds == labels).mean() @dataclass class _snake_case : __A : str =field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"}) __A : Optional[str] =field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Pretrained config name or path if not the same as model_name"}) __A : Optional[str] =field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) __A : Optional[str] =field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class _snake_case : __A : str =field(metadata={"help": "The name of the task to train on: " + ", ".join(processors.keys())}) __A : str =field(metadata={"help": "Should contain the data files for the task."}) __A : int =field( default=1_28 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) __A : bool =field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a__ ( ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( "Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("Training/evaluation parameters %s" , _SCREAMING_SNAKE_CASE ) # Set seed set_seed(training_args.seed ) try: UpperCAmelCase_ : Union[str, Any] = processors[data_args.task_name]() UpperCAmelCase_ : int = processor.get_labels() UpperCAmelCase_ : int = len(_SCREAMING_SNAKE_CASE ) except KeyError: raise ValueError("Task not found: %s" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase_ : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=_SCREAMING_SNAKE_CASE , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) UpperCAmelCase_ : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) UpperCAmelCase_ : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=_SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) # Get datasets UpperCAmelCase_ : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_SCREAMING_SNAKE_CASE , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) UpperCAmelCase_ : List[str] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_SCREAMING_SNAKE_CASE , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(_SCREAMING_SNAKE_CASE : EvalPrediction ) -> Dict: UpperCAmelCase_ : List[Any] = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE , p.label_ids )} # Data collator UpperCAmelCase_ : str = DataCollatorWithPadding(_SCREAMING_SNAKE_CASE , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer UpperCAmelCase_ : Union[str, Any] = Trainer( model=_SCREAMING_SNAKE_CASE , args=_SCREAMING_SNAKE_CASE , train_dataset=_SCREAMING_SNAKE_CASE , eval_dataset=_SCREAMING_SNAKE_CASE , compute_metrics=_SCREAMING_SNAKE_CASE , data_collator=_SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCAmelCase_ : List[Any] = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) UpperCAmelCase_ : List[Any] = trainer.evaluate() UpperCAmelCase_ : List[Any] = os.path.join(training_args.output_dir , "eval_results.txt" ) if trainer.is_world_master(): with open(_SCREAMING_SNAKE_CASE , "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(" %s = %s" , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) writer.write("%s = %s\n" % (key, value) ) results.update(_SCREAMING_SNAKE_CASE ) return results def a__ ( _SCREAMING_SNAKE_CASE : Optional[int] ) -> int: """simple docstring""" main() if __name__ == "__main__": main()
67
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return int((input_a, input_a).count(0 ) != 0 ) def a__ ( ) -> None: """simple docstring""" assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
67
1
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def UpperCamelCase_( lowerCamelCase_="" ) -> str: _lowercase : Optional[Any] = tempfile.mkdtemp() return os.path.join(lowerCamelCase_ , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[Any] = torch.rand(12, dtype=torch.floataa) - 0.5 _lowercase : Dict = AgentAudio(lowerCamelCase) _lowercase : List[Any] = str(agent_type.to_string()) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCamelCase, agent_type.to_raw(), atol=1E-4)) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowerCamelCase)) # Ensure that the file contains the same value as the original tensor _lowercase , _lowercase : List[Any] = sf.read(lowerCamelCase) self.assertTrue(torch.allclose(lowerCamelCase, torch.tensor(lowerCamelCase), atol=1E-4)) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = torch.rand(12, dtype=torch.floataa) - 0.5 _lowercase : Optional[Any] = get_new_path(suffix='.wav') sf.write(lowerCamelCase, lowerCamelCase, 1_60_00) _lowercase : List[str] = AgentAudio(lowerCamelCase) self.assertTrue(torch.allclose(lowerCamelCase, agent_type.to_raw(), atol=1E-4)) self.assertEqual(agent_type.to_string(), lowerCamelCase) @require_vision @require_torch class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : str = torch.randint(0, 2_56, (64, 64, 3)) _lowercase : List[Any] = AgentImage(lowerCamelCase) _lowercase : Any = str(agent_type.to_string()) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCamelCase, agent_type._tensor, atol=1E-4)) self.assertIsInstance(agent_type.to_raw(), Image.Image) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCamelCase)) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = Path(get_tests_dir('fixtures/tests_samples/COCO')) / '000000039769.png' _lowercase : Any = Image.open(lowerCamelCase) _lowercase : Optional[int] = AgentImage(lowerCamelCase) self.assertTrue(path.samefile(agent_type.to_string())) self.assertTrue(image == agent_type.to_raw()) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCamelCase)) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Optional[int] = Path(get_tests_dir('fixtures/tests_samples/COCO')) / '000000039769.png' _lowercase : Optional[int] = Image.open(lowerCamelCase) _lowercase : List[str] = AgentImage(lowerCamelCase) self.assertFalse(path.samefile(agent_type.to_string())) self.assertTrue(image == agent_type.to_raw()) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCamelCase)) class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Optional[int] = 'Hey!' _lowercase : List[str] = AgentText(lowerCamelCase) self.assertEqual(lowerCamelCase, agent_type.to_string()) self.assertEqual(lowerCamelCase, agent_type.to_raw()) self.assertEqual(lowerCamelCase, lowerCamelCase)
21
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> float: if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) __lowerCamelCase : int = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(lowerCamelCase__ ) ) return round(lowerCamelCase__ , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
73
0
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class UpperCAmelCase (unittest.TestCase ): """simple docstring""" def _snake_case ( self ): lowercase__ = '''ylacombe/bark-small''' lowercase__ = tempfile.mkdtemp() lowercase__ = '''en_speaker_1''' lowercase__ = '''This is a test string''' lowercase__ = '''speaker_embeddings_path.json''' lowercase__ = '''speaker_embeddings''' def _snake_case ( self , **_UpperCAmelCase ): return AutoTokenizer.from_pretrained(self.checkpoint , **_UpperCAmelCase ) def _snake_case ( self ): shutil.rmtree(self.tmpdirname ) def _snake_case ( self ): lowercase__ = self.get_tokenizer() lowercase__ = BarkProcessor(tokenizer=_UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) lowercase__ = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def _snake_case ( self ): lowercase__ = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) lowercase__ = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) lowercase__ = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token='''(BOS)''' , eos_token='''(EOS)''' , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def _snake_case ( self ): lowercase__ = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) lowercase__ = 35 lowercase__ = 2 lowercase__ = 8 lowercase__ = { '''semantic_prompt''': np.ones(_UpperCAmelCase ), '''coarse_prompt''': np.ones((nb_codebooks_coarse, seq_len) ), '''fine_prompt''': np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset lowercase__ = processor(text=self.input_string , voice_preset=_UpperCAmelCase ) lowercase__ = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(_UpperCAmelCase , np.array([] ) ).tolist() ) # test loading voice preset from npz file lowercase__ = os.path.join(self.tmpdirname , '''file.npz''' ) np.savez(_UpperCAmelCase , **_UpperCAmelCase ) lowercase__ = processor(text=self.input_string , voice_preset=_UpperCAmelCase ) lowercase__ = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(_UpperCAmelCase , np.array([] ) ).tolist() ) # test loading voice preset from the hub lowercase__ = processor(text=self.input_string , voice_preset=self.voice_preset ) def _snake_case ( self ): lowercase__ = self.get_tokenizer() lowercase__ = BarkProcessor(tokenizer=_UpperCAmelCase ) lowercase__ = processor(text=self.input_string ) lowercase__ = tokenizer( self.input_string , padding='''max_length''' , max_length=256 , add_special_tokens=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
360
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) __A = "hf-internal-testing/tiny-random-bert" __A = os.path.join(TRANSFORMERS_CACHE, "models--hf-internal-testing--tiny-random-bert") __A = "9b8c223d42b2188cb49d29af482996f9d0f3e5a6" class UpperCAmelCase (unittest.TestCase ): """simple docstring""" def _snake_case ( self ): lowercase__: Union[str, Any] = cached_file(_UpperCAmelCase , _UpperCAmelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(_UpperCAmelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) ) with open(os.path.join(_UpperCAmelCase , '''refs''' , '''main''' ) ) as f: lowercase__: Dict = f.read() self.assertEqual(_UpperCAmelCase , os.path.join(_UpperCAmelCase , '''snapshots''' , _UpperCAmelCase , _UpperCAmelCase ) ) self.assertTrue(os.path.isfile(_UpperCAmelCase ) ) # File is cached at the same place the second time. lowercase__: Any = cached_file(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) # Using a specific revision to test the full commit hash. lowercase__: Dict = cached_file(_UpperCAmelCase , _UpperCAmelCase , revision='''9b8c223''' ) self.assertEqual(_UpperCAmelCase , os.path.join(_UpperCAmelCase , '''snapshots''' , _UpperCAmelCase , _UpperCAmelCase ) ) def _snake_case ( self ): with self.assertRaisesRegex(_UpperCAmelCase , '''is not a valid model identifier''' ): lowercase__: int = cached_file('''tiny-random-bert''' , _UpperCAmelCase ) with self.assertRaisesRegex(_UpperCAmelCase , '''is not a valid git identifier''' ): lowercase__: List[Any] = cached_file(_UpperCAmelCase , _UpperCAmelCase , revision='''aaaa''' ) with self.assertRaisesRegex(_UpperCAmelCase , '''does not appear to have a file named''' ): lowercase__: Dict = cached_file(_UpperCAmelCase , '''conf''' ) def _snake_case ( self ): with self.assertRaisesRegex(_UpperCAmelCase , '''does not appear to have a file named''' ): lowercase__: Optional[Any] = cached_file(_UpperCAmelCase , '''conf''' ) with open(os.path.join(_UpperCAmelCase , '''refs''' , '''main''' ) ) as f: lowercase__: int = f.read() self.assertTrue(os.path.isfile(os.path.join(_UpperCAmelCase , '''.no_exist''' , _UpperCAmelCase , '''conf''' ) ) ) lowercase__: Dict = cached_file(_UpperCAmelCase , '''conf''' , _raise_exceptions_for_missing_entries=_UpperCAmelCase ) self.assertIsNone(_UpperCAmelCase ) lowercase__: List[str] = cached_file(_UpperCAmelCase , '''conf''' , local_files_only=_UpperCAmelCase , _raise_exceptions_for_missing_entries=_UpperCAmelCase ) self.assertIsNone(_UpperCAmelCase ) lowercase__: Union[str, Any] = mock.Mock() lowercase__: str = 500 lowercase__: Union[str, Any] = {} lowercase__: List[str] = HTTPError lowercase__: int = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=_UpperCAmelCase ) as mock_head: lowercase__: Any = cached_file(_UpperCAmelCase , '''conf''' , _raise_exceptions_for_connection_errors=_UpperCAmelCase ) self.assertIsNone(_UpperCAmelCase ) # This check we did call the fake head request mock_head.assert_called() def _snake_case ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , _UpperCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , _UpperCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , _UpperCAmelCase ) ) def _snake_case ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(_UpperCAmelCase , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , _UpperCAmelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(_UpperCAmelCase , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , _UpperCAmelCase , revision='''ahaha''' ) lowercase__: Optional[Any] = get_file_from_repo('''bert-base-cased''' , _UpperCAmelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. lowercase__: Optional[Any] = json.loads(open(_UpperCAmelCase , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 768 ) def _snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: lowercase__: Any = Path(_UpperCAmelCase ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(_UpperCAmelCase , '''a.txt''' ) , str(_UpperCAmelCase ) ) self.assertIsNone(get_file_from_repo(_UpperCAmelCase , '''b.txt''' ) )
2
0
'''simple docstring''' import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = (DDIMParallelScheduler,) lowerCAmelCase = (('''eta''', 0.0), ('''num_inference_steps''', 50)) def SCREAMING_SNAKE_CASE ( self , **_UpperCAmelCase): '''simple docstring''' __A : List[Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'clip_sample': True, } config.update(**_UpperCAmelCase) return config def SCREAMING_SNAKE_CASE ( self , **_UpperCAmelCase): '''simple docstring''' __A : Tuple = self.scheduler_classes[0] __A : str = self.get_scheduler_config(**_UpperCAmelCase) __A : int = scheduler_class(**_UpperCAmelCase) __A ,__A : Optional[int] = 10, 0.0 __A : str = self.dummy_model() __A : Any = self.dummy_sample_deter scheduler.set_timesteps(_UpperCAmelCase) for t in scheduler.timesteps: __A : List[Any] = model(_UpperCAmelCase , _UpperCAmelCase) __A : Union[str, Any] = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase).prev_sample return sample def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=_UpperCAmelCase) __A : Tuple = self.scheduler_classes[0] __A : int = self.get_scheduler_config(steps_offset=1) __A : List[str] = scheduler_class(**_UpperCAmelCase) scheduler.set_timesteps(5) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1])) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2]): self.check_over_configs(beta_start=_UpperCAmelCase , beta_end=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' self.check_over_configs(thresholding=_UpperCAmelCase) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=_UpperCAmelCase , prediction_type=_UpperCAmelCase , sample_max_value=_UpperCAmelCase , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for t in [1, 10, 49]: self.check_over_forward(time_step=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500]): self.check_over_forward(time_step=_UpperCAmelCase , num_inference_steps=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0]): self.check_over_forward(time_step=_UpperCAmelCase , eta=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Any = self.scheduler_classes[0] __A : List[str] = self.get_scheduler_config() __A : Optional[int] = scheduler_class(**_UpperCAmelCase) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0) - 0.0)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400) - 0.14771)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960) - 0.32460)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0) - 0.0)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486) - 0.00979)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998) - 0.02)) < 1e-5 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = self.scheduler_classes[0] __A : Optional[int] = self.get_scheduler_config() __A : Optional[Any] = scheduler_class(**_UpperCAmelCase) __A ,__A : Union[str, Any] = 10, 0.0 scheduler.set_timesteps(_UpperCAmelCase) __A : int = self.dummy_model() __A : Tuple = self.dummy_sample_deter __A : Optional[Any] = self.dummy_sample_deter + 0.1 __A : Union[str, Any] = self.dummy_sample_deter - 0.1 __A : str = samplea.shape[0] __A : Optional[Any] = torch.stack([samplea, samplea, samplea] , dim=0) __A : Union[str, Any] = torch.arange(_UpperCAmelCase)[0:3, None].repeat(1 , _UpperCAmelCase) __A : int = model(samples.flatten(0 , 1) , timesteps.flatten(0 , 1)) __A : Dict = scheduler.batch_step_no_noise(_UpperCAmelCase , timesteps.flatten(0 , 1) , samples.flatten(0 , 1) , _UpperCAmelCase) __A : List[Any] = torch.sum(torch.abs(_UpperCAmelCase)) __A : Any = torch.mean(torch.abs(_UpperCAmelCase)) assert abs(result_sum.item() - 1147.7904) < 1e-2 assert abs(result_mean.item() - 0.4982) < 1e-3 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = self.full_loop() __A : str = torch.sum(torch.abs(_UpperCAmelCase)) __A : int = torch.mean(torch.abs(_UpperCAmelCase)) assert abs(result_sum.item() - 172.0067) < 1e-2 assert abs(result_mean.item() - 0.223967) < 1e-3 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.full_loop(prediction_type='v_prediction') __A : Union[str, Any] = torch.sum(torch.abs(_UpperCAmelCase)) __A : Dict = torch.mean(torch.abs(_UpperCAmelCase)) assert abs(result_sum.item() - 52.5302) < 1e-2 assert abs(result_mean.item() - 0.0684) < 1e-3 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01) __A : Optional[Any] = torch.sum(torch.abs(_UpperCAmelCase)) __A : List[Any] = torch.mean(torch.abs(_UpperCAmelCase)) assert abs(result_sum.item() - 149.8295) < 1e-2 assert abs(result_mean.item() - 0.1951) < 1e-3 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01) __A : Dict = torch.sum(torch.abs(_UpperCAmelCase)) __A : Any = torch.mean(torch.abs(_UpperCAmelCase)) assert abs(result_sum.item() - 149.0784) < 1e-2 assert abs(result_mean.item() - 0.1941) < 1e-3
190
'''simple docstring''' import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowercase__ : List[Any] = pd.read_csv('''sample_data.csv''', header=None) lowercase__ : Union[str, Any] = df.shape[:1][0] # If you're using some other dataset input the target column lowercase__ : Any = df.iloc[:, 1:2] lowercase__ : int = actual_data.values.reshape(len_data, 1) lowercase__ : int = MinMaxScaler().fit_transform(actual_data) lowercase__ : Dict = 10 lowercase__ : List[str] = 5 lowercase__ : Dict = 20 lowercase__ : Dict = len_data - periods * look_back lowercase__ : Any = actual_data[:division] lowercase__ : Optional[int] = actual_data[division - look_back :] lowercase__ , lowercase__ : Optional[Any] = [], [] lowercase__ , lowercase__ : Tuple = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowercase__ : List[Any] = np.array(train_x) lowercase__ : List[Any] = np.array(test_x) lowercase__ : str = np.array([list(i.ravel()) for i in train_y]) lowercase__ : Any = np.array([list(i.ravel()) for i in test_y]) lowercase__ : Optional[Any] = Sequential() model.add(LSTM(1_28, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(1_28, 1))) model.add(Dense(forward_days)) model.compile(loss='''mean_squared_error''', optimizer='''adam''') lowercase__ : Tuple = model.fit( x_train, y_train, epochs=1_50, verbose=1, shuffle=True, batch_size=4 ) lowercase__ : int = model.predict(x_test)
190
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __UpperCAmelCase ( _lowerCamelCase , unittest.TestCase ): __lowercase = KandinskyVaaControlnetPipeline __lowercase = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowercase = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowercase = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __lowercase = False @property def lowerCamelCase ( self ): """simple docstring""" return 32 @property def lowerCamelCase ( self ): """simple docstring""" return 32 @property def lowerCamelCase ( self ): """simple docstring""" return self.time_input_dim @property def lowerCamelCase ( self ): """simple docstring""" return self.time_input_dim * 4 @property def lowerCamelCase ( self ): """simple docstring""" return 1_00 @property def lowerCamelCase ( self ): """simple docstring""" torch.manual_seed(0 ) _snake_case = { 'in_channels': 8, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image_hint', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _snake_case = UNetaDConditionModel(**lowerCAmelCase_ ) return model @property def lowerCamelCase ( self ): """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def lowerCamelCase ( self ): """simple docstring""" torch.manual_seed(0 ) _snake_case = VQModel(**self.dummy_movq_kwargs ) return model def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.dummy_unet _snake_case = self.dummy_movq _snake_case = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowerCAmelCase_ , set_alpha_to_one=lowerCAmelCase_ , steps_offset=1 , prediction_type='epsilon' , thresholding=lowerCAmelCase_ , ) _snake_case = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=0 ): """simple docstring""" _snake_case = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) _snake_case = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( lowerCAmelCase_ ) # create hint _snake_case = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) if str(lowerCAmelCase_ ).startswith('mps' ): _snake_case = torch.manual_seed(lowerCAmelCase_ ) else: _snake_case = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _snake_case = { 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'hint': hint, 'generator': generator, 'height': 64, 'width': 64, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'cpu' _snake_case = self.get_dummy_components() _snake_case = self.pipeline_class(**lowerCAmelCase_ ) _snake_case = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = pipe(**self.get_dummy_inputs(lowerCAmelCase_ ) ) _snake_case = output.images _snake_case = pipe( **self.get_dummy_inputs(lowerCAmelCase_ ) , return_dict=lowerCAmelCase_ , )[0] _snake_case = image[0, -3:, -3:, -1] _snake_case = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _snake_case = np.array( [0.6959826, 0.868279, 0.7558092, 0.68769467, 0.85805804, 0.65977496, 0.44885302, 0.5959111, 0.4251595] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy' ) _snake_case = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/hint_image_cat.png' ) _snake_case = torch.from_numpy(np.array(lowerCAmelCase_ ) ).float() / 255.0 _snake_case = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) _snake_case = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(lowerCAmelCase_ ) _snake_case = KandinskyVaaControlnetPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-controlnet-depth' , torch_dtype=torch.floataa ) _snake_case = pipeline.to(lowerCAmelCase_ ) pipeline.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A robot, 4k photo' _snake_case = torch.Generator(device='cuda' ).manual_seed(0 ) _snake_case , _snake_case = pipe_prior( lowerCAmelCase_ , generator=lowerCAmelCase_ , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _snake_case = torch.Generator(device='cuda' ).manual_seed(0 ) _snake_case = pipeline( image_embeds=lowerCAmelCase_ , negative_image_embeds=lowerCAmelCase_ , hint=lowerCAmelCase_ , generator=lowerCAmelCase_ , num_inference_steps=1_00 , output_type='np' , ) _snake_case = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(lowerCAmelCase_ , lowerCAmelCase_ )
160
'''simple docstring''' import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowercase : int = NewType("DataClass", Any) lowercase : Dict = NewType("DataClassType", Any) def SCREAMING_SNAKE_CASE__ ( __A ) -> Optional[Any]: if isinstance(__A , __A ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).' ) def SCREAMING_SNAKE_CASE__ ( __A ) -> Callable[[str], Any]: _snake_case = {str(__A ): choice for choice in choices} return lambda __A : str_to_choice.get(__A , __A ) def SCREAMING_SNAKE_CASE__ ( *, __A = None , __A = None , __A = dataclasses.MISSING , __A = dataclasses.MISSING , __A = None , **__A , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _snake_case = {} if aliases is not None: _snake_case = aliases if help is not None: _snake_case = help return dataclasses.field(metadata=__A , default=__A , default_factory=__A , **__A ) class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 def __init__( self , lowerCAmelCase_ , **lowerCAmelCase_ ): """simple docstring""" if "formatter_class" not in kwargs: _snake_case = ArgumentDefaultsHelpFormatter super().__init__(**lowerCAmelCase_ ) if dataclasses.is_dataclass(lowerCAmelCase_ ): _snake_case = [dataclass_types] _snake_case = list(lowerCAmelCase_ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = F'--{field.name}' _snake_case = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , lowerCAmelCase_ ): raise RuntimeError( 'Unresolved type detected, which should have been done with the help of ' '`typing.get_type_hints` method by default' ) _snake_case = kwargs.pop('aliases' , [] ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _snake_case = [aliases] _snake_case = getattr(field.type , '__origin__' , field.type ) if origin_type is Union or (hasattr(lowerCAmelCase_ , 'UnionType' ) and isinstance(lowerCAmelCase_ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowerCAmelCase_ ) not in field.type.__args__ ): raise ValueError( 'Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because' ' the argument parser only supports one type per argument.' F' Problem encountered in field \'{field.name}\'.' ) if type(lowerCAmelCase_ ) not in field.type.__args__: # filter `str` in Union _snake_case = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _snake_case = getattr(field.type , '__origin__' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _snake_case = ( field.type.__args__[0] if isinstance(lowerCAmelCase_ , field.type.__args__[1] ) else field.type.__args__[1] ) _snake_case = getattr(field.type , '__origin__' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _snake_case = {} if origin_type is Literal or (isinstance(field.type , lowerCAmelCase_ ) and issubclass(field.type , lowerCAmelCase_ )): if origin_type is Literal: _snake_case = field.type.__args__ else: _snake_case = [x.value for x in field.type] _snake_case = make_choice_type_function(kwargs['choices'] ) if field.default is not dataclasses.MISSING: _snake_case = field.default else: _snake_case = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _snake_case = copy(lowerCAmelCase_ ) # Hack because type=bool in argparse does not behave as we want. _snake_case = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _snake_case = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _snake_case = default # This tells argparse we accept 0 or 1 value after --field_name _snake_case = '?' # This is the value that will get picked if we do --field_name (without value) _snake_case = True elif isclass(lowerCAmelCase_ ) and issubclass(lowerCAmelCase_ , lowerCAmelCase_ ): _snake_case = field.type.__args__[0] _snake_case = '+' if field.default_factory is not dataclasses.MISSING: _snake_case = field.default_factory() elif field.default is dataclasses.MISSING: _snake_case = True else: _snake_case = field.type if field.default is not dataclasses.MISSING: _snake_case = field.default elif field.default_factory is not dataclasses.MISSING: _snake_case = field.default_factory() else: _snake_case = True parser.add_argument(lowerCAmelCase_ , *lowerCAmelCase_ , **lowerCAmelCase_ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _snake_case = False parser.add_argument(F'--no_{field.name}' , action='store_false' , dest=field.name , **lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if hasattr(lowerCAmelCase_ , '_argument_group_name' ): _snake_case = self.add_argument_group(dtype._argument_group_name ) else: _snake_case = self try: _snake_case = get_type_hints(lowerCAmelCase_ ) except NameError: raise RuntimeError( F'Type resolution failed for {dtype}. Try declaring the class in global scope or ' 'removing line of `from __future__ import annotations` which opts in Postponed ' 'Evaluation of Annotations (PEP 563)' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(lowerCAmelCase_ ): _snake_case = '.'.join(map(lowerCAmelCase_ , sys.version_info[:3] ) ) raise RuntimeError( F'Type resolution failed for {dtype} on Python {python_version}. Try removing ' 'line of `from __future__ import annotations` which opts in union types as ' '`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ' 'support Python versions that lower than 3.10, you need to use ' '`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ' '`X | None`.' ) from ex raise for field in dataclasses.fields(lowerCAmelCase_ ): if not field.init: continue _snake_case = type_hints[field.name] self._parse_dataclass_field(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_=None , lowerCAmelCase_=None , ): """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _snake_case = [] if args_filename: args_files.append(Path(lowerCAmelCase_ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('.args' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _snake_case = ArgumentParser() args_file_parser.add_argument(lowerCAmelCase_ , type=lowerCAmelCase_ , action='append' ) # Use only remaining args for further parsing (remove the args_file_flag) _snake_case , _snake_case = args_file_parser.parse_known_args(args=lowerCAmelCase_ ) _snake_case = vars(lowerCAmelCase_ ).get(args_file_flag.lstrip('-' ) , lowerCAmelCase_ ) if cmd_args_file_paths: args_files.extend([Path(lowerCAmelCase_ ) for p in cmd_args_file_paths] ) _snake_case = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _snake_case = file_args + args if args is not None else file_args + sys.argv[1:] _snake_case , _snake_case = self.parse_known_args(args=lowerCAmelCase_ ) _snake_case = [] for dtype in self.dataclass_types: _snake_case = {f.name for f in dataclasses.fields(lowerCAmelCase_ ) if f.init} _snake_case = {k: v for k, v in vars(lowerCAmelCase_ ).items() if k in keys} for k in keys: delattr(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = dtype(**lowerCAmelCase_ ) outputs.append(lowerCAmelCase_ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowerCAmelCase_ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F'Some specified arguments are not used by the HfArgumentParser: {remaining_args}' ) return (*outputs,) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = False ): """simple docstring""" _snake_case = set(args.keys() ) _snake_case = [] for dtype in self.dataclass_types: _snake_case = {f.name for f in dataclasses.fields(lowerCAmelCase_ ) if f.init} _snake_case = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _snake_case = dtype(**lowerCAmelCase_ ) outputs.append(lowerCAmelCase_ ) if not allow_extra_keys and unused_keys: raise ValueError(F'Some keys are not used by the HfArgumentParser: {sorted(lowerCAmelCase_ )}' ) return tuple(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = False ): """simple docstring""" with open(Path(lowerCAmelCase_ ) , encoding='utf-8' ) as open_json_file: _snake_case = json.loads(open_json_file.read() ) _snake_case = self.parse_dict(lowerCAmelCase_ , allow_extra_keys=lowerCAmelCase_ ) return tuple(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = False ): """simple docstring""" _snake_case = self.parse_dict(yaml.safe_load(Path(lowerCAmelCase_ ).read_text() ) , allow_extra_keys=lowerCAmelCase_ ) return tuple(lowerCAmelCase_ )
160
1
"""simple docstring""" import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def _snake_case ( snake_case__ : str , snake_case__ : str ): A = RobertaPreLayerNormConfig.from_pretrained( snake_case__ , architectures=['RobertaPreLayerNormForMaskedLM'] ) # convert state_dict A = torch.load(hf_hub_download(repo_id=snake_case__ , filename='pytorch_model.bin' ) ) A = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith('roberta.' ): A = 'roberta_prelayernorm.' + tensor_key[len('roberta.' ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith('.self.LayerNorm.weight' ) or tensor_key.endswith('.self.LayerNorm.bias' ): continue A = tensor_value A = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=snake_case__ , config=snake_case__ , state_dict=snake_case__ ) model.save_pretrained(snake_case__ ) # convert tokenizer A = AutoTokenizer.from_pretrained(snake_case__ ) tokenizer.save_pretrained(snake_case__ ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint-repo''', default=None, type=str, required=True, help='''Path the official PyTorch dump, e.g. \'andreasmadsen/efficient_mlm_m0.40\'.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _lowercase = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
74
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = 1 _lowercase : Any = 3 _lowercase : Tuple = (32, 32) _lowercase : Tuple = floats_tensor((batch_size, num_channels) + sizes, rng=random.Random(0)).to(lowerCamelCase) return image @property def UpperCamelCase ( self) -> str: """simple docstring""" torch.manual_seed(0) _lowercase : Dict = UNetaDConditionModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=4, out_channels=4, down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D'), up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D'), cross_attention_dim=32, ) return model @property def UpperCamelCase ( self) -> List[Any]: """simple docstring""" torch.manual_seed(0) _lowercase : str = AutoencoderKL( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'], up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'], latent_channels=4, ) return model @property def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" torch.manual_seed(0) _lowercase : Optional[int] = RobertaSeriesConfig( hidden_size=32, project_dim=32, intermediate_size=37, layer_norm_eps=1E-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=50_06, ) return RobertaSeriesModelWithTransformation(lowerCamelCase) @property def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" def extract(*lowerCamelCase, **lowerCamelCase): class _lowerCamelCase: def __init__( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = torch.ones([0]) def UpperCamelCase ( self, lowerCamelCase) -> int: """simple docstring""" self.pixel_values.to(lowerCamelCase) return self return Out() return extract def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : List[Any] = self.dummy_cond_unet _lowercase : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCamelCase) _lowercase : Optional[Any] = self.dummy_vae _lowercase : List[Any] = self.dummy_text_encoder _lowercase : Any = XLMRobertaTokenizer.from_pretrained('hf-internal-testing/tiny-xlm-roberta') _lowercase : Tuple = 77 _lowercase : int = self.dummy_image.to(lowerCamelCase) _lowercase : int = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk _lowercase : Union[str, Any] = AltDiffusionImgaImgPipeline( unet=lowerCamelCase, scheduler=lowerCamelCase, vae=lowerCamelCase, text_encoder=lowerCamelCase, tokenizer=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=self.dummy_extractor, ) _lowercase : List[Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor, do_normalize=lowerCamelCase) _lowercase : Optional[int] = alt_pipe.to(lowerCamelCase) alt_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = 'A painting of a squirrel eating a burger' _lowercase : Dict = torch.Generator(device=lowerCamelCase).manual_seed(0) _lowercase : Any = alt_pipe( [prompt], generator=lowerCamelCase, guidance_scale=6.0, num_inference_steps=2, output_type='np', image=lowerCamelCase, ) _lowercase : Optional[int] = output.images _lowercase : Optional[Any] = torch.Generator(device=lowerCamelCase).manual_seed(0) _lowercase : Optional[Any] = alt_pipe( [prompt], generator=lowerCamelCase, guidance_scale=6.0, num_inference_steps=2, output_type='np', image=lowerCamelCase, return_dict=lowerCamelCase, )[0] _lowercase : Optional[int] = image[0, -3:, -3:, -1] _lowercase : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowercase : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 5E-3 @unittest.skipIf(torch_device != 'cuda', 'This test requires a GPU') def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[Any] = self.dummy_cond_unet _lowercase : Tuple = PNDMScheduler(skip_prk_steps=lowerCamelCase) _lowercase : str = self.dummy_vae _lowercase : Optional[Any] = self.dummy_text_encoder _lowercase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('hf-internal-testing/tiny-xlm-roberta') _lowercase : Optional[Any] = 77 _lowercase : str = self.dummy_image.to(lowerCamelCase) # put models in fp16 _lowercase : List[str] = unet.half() _lowercase : List[Any] = vae.half() _lowercase : Any = bert.half() # make sure here that pndm scheduler skips prk _lowercase : Union[str, Any] = AltDiffusionImgaImgPipeline( unet=lowerCamelCase, scheduler=lowerCamelCase, vae=lowerCamelCase, text_encoder=lowerCamelCase, tokenizer=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=self.dummy_extractor, ) _lowercase : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor, do_normalize=lowerCamelCase) _lowercase : Any = alt_pipe.to(lowerCamelCase) alt_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : int = 'A painting of a squirrel eating a burger' _lowercase : Optional[Any] = torch.manual_seed(0) _lowercase : Union[str, Any] = alt_pipe( [prompt], generator=lowerCamelCase, num_inference_steps=2, output_type='np', image=lowerCamelCase, ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != 'cuda', 'This test requires a GPU') def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') # resize to resolution that is divisible by 8 but not 16 or 32 _lowercase : str = init_image.resize((7_60, 5_04)) _lowercase : Optional[int] = 'BAAI/AltDiffusion' _lowercase : str = AltDiffusionImgaImgPipeline.from_pretrained( lowerCamelCase, safety_checker=lowerCamelCase, ) pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) pipe.enable_attention_slicing() _lowercase : List[str] = 'A fantasy landscape, trending on artstation' _lowercase : Any = torch.manual_seed(0) _lowercase : Dict = pipe( prompt=lowerCamelCase, image=lowerCamelCase, strength=0.7_5, guidance_scale=7.5, generator=lowerCamelCase, output_type='np', ) _lowercase : List[str] = output.images[0] _lowercase : Tuple = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 7_60, 3) _lowercase : Optional[Any] = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') _lowercase : str = init_image.resize((7_68, 5_12)) _lowercase : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy') _lowercase : str = 'BAAI/AltDiffusion' _lowercase : Optional[Any] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCamelCase, safety_checker=lowerCamelCase, ) pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) pipe.enable_attention_slicing() _lowercase : int = 'A fantasy landscape, trending on artstation' _lowercase : List[Any] = torch.manual_seed(0) _lowercase : int = pipe( prompt=lowerCamelCase, image=lowerCamelCase, strength=0.7_5, guidance_scale=7.5, generator=lowerCamelCase, output_type='np', ) _lowercase : Union[str, Any] = output.images[0] assert image.shape == (5_12, 7_68, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image).max() < 1E-2
21
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __magic_name__ : def __init__( self , __snake_case , __snake_case=3 , __snake_case=32 , __snake_case=3 , __snake_case=10 , __snake_case=[10, 20, 30, 40] , __snake_case=[1, 1, 2, 1] , __snake_case=True , __snake_case=True , __snake_case="relu" , __snake_case=3 , __snake_case=None , ) -> int: '''simple docstring''' __a =parent __a =batch_size __a =image_size __a =num_channels __a =embeddings_size __a =hidden_sizes __a =depths __a =is_training __a =use_labels __a =hidden_act __a =num_labels __a =scope __a =len(_a ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' __a =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a =None if self.use_labels: __a =ids_tensor([self.batch_size] , self.num_labels ) __a =self.get_config() return config, pixel_values, labels def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> List[str]: '''simple docstring''' __a =TFResNetModel(config=_a ) __a =model(_a ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> Union[str, Any]: '''simple docstring''' __a =self.num_labels __a =TFResNetForImageClassification(_a ) __a =model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __magic_name__ ( self ) -> str: '''simple docstring''' __a =self.prepare_config_and_inputs() __a , __a , __a =config_and_inputs __a ={'pixel_values': pixel_values} return config, inputs_dict @require_tf class __magic_name__ ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): SCREAMING_SNAKE_CASE = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE = ( {'''feature-extraction''': TFResNetModel, '''image-classification''': TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def __magic_name__ ( self ) -> Tuple: '''simple docstring''' __a =TFResNetModelTester(self ) __a =ConfigTester(self , config_class=_a , has_text_modality=_a ) def __magic_name__ ( self ) -> Any: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __magic_name__ ( self ) -> List[str]: '''simple docstring''' return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def __magic_name__ ( self ) -> int: '''simple docstring''' pass def __magic_name__ ( self ) -> int: '''simple docstring''' __a , __a =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a =model_class(_a ) __a =inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a =[*signature.parameters.keys()] __a =['pixel_values'] self.assertListEqual(arg_names[:1] , _a ) def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __magic_name__ ( self ) -> Tuple: '''simple docstring''' def check_hidden_states_output(__snake_case , __snake_case , __snake_case ): __a =model_class(_a ) __a =model(**self._prepare_for_class(_a , _a ) ) __a =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __a =self.model_tester.num_stages self.assertEqual(len(_a ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __a , __a =self.model_tester.prepare_config_and_inputs_for_common() __a =['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: __a =layer_type __a =True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a =True check_hidden_states_output(_a , _a , _a ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def __magic_name__ ( self ) -> Any: '''simple docstring''' for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a =TFResNetModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def UpperCamelCase_( ): """simple docstring""" __a =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __magic_name__ ( unittest.TestCase ): @cached_property def __magic_name__ ( self ) -> Any: '''simple docstring''' return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a =TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __a =self.default_image_processor __a =prepare_img() __a =image_processor(images=_a , return_tensors='tf' ) # forward pass __a =model(**_a ) # verify the logits __a =tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _a ) __a =tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _a , atol=1e-4 ) )
363
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase : List[str] = logging.get_logger(__name__) _lowerCAmelCase : List[Any] = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'yolos' def __init__( self , __snake_case=768 , __snake_case=12 , __snake_case=12 , __snake_case=3072 , __snake_case="gelu" , __snake_case=0.0 , __snake_case=0.0 , __snake_case=0.02 , __snake_case=1e-12 , __snake_case=[512, 864] , __snake_case=16 , __snake_case=3 , __snake_case=True , __snake_case=100 , __snake_case=True , __snake_case=False , __snake_case=1 , __snake_case=5 , __snake_case=2 , __snake_case=5 , __snake_case=2 , __snake_case=0.1 , **__snake_case , ) -> str: '''simple docstring''' super().__init__(**__snake_case ) __a =hidden_size __a =num_hidden_layers __a =num_attention_heads __a =intermediate_size __a =hidden_act __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =initializer_range __a =layer_norm_eps __a =image_size __a =patch_size __a =num_channels __a =qkv_bias __a =num_detection_tokens __a =use_mid_position_embeddings __a =auxiliary_loss # Hungarian matcher __a =class_cost __a =bbox_cost __a =giou_cost # Loss coefficients __a =bbox_loss_coefficient __a =giou_loss_coefficient __a =eos_coefficient class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = version.parse('1.11' ) @property def __magic_name__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def __magic_name__ ( self ) -> float: '''simple docstring''' return 1e-4 @property def __magic_name__ ( self ) -> int: '''simple docstring''' return 12
308
0
"""simple docstring""" import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline lowerCAmelCase_ = datasets.utils.logging.get_logger(__name__) @dataclass class __A ( datasets.BuilderConfig ): '''simple docstring''' lowerCAmelCase : Optional[datasets.Features] = None lowerCAmelCase : str = "utf-8" lowerCAmelCase : Optional[str] = None lowerCAmelCase : Optional[str] = None lowerCAmelCase : bool = True # deprecated lowerCAmelCase : Optional[int] = None # deprecated lowerCAmelCase : int = 1_0 << 2_0 # 10MB lowerCAmelCase : Optional[bool] = None class __A ( datasets.ArrowBasedBuilder ): '''simple docstring''' lowerCAmelCase : int = JsonConfig def UpperCAmelCase ( self : int ) -> int: """simple docstring""" if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowercase__ : int = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def UpperCAmelCase ( self : Dict ,_snake_case : Optional[int] ) -> Dict: """simple docstring""" if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowercase__ : int = dl_manager.download_and_extract(self.config.data_files ) if isinstance(_snake_case ,(str, list, tuple) ): lowercase__ : Any = data_files if isinstance(_snake_case ,_snake_case ): lowercase__ : Optional[Any] = [files] lowercase__ : int = [dl_manager.iter_files(_snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN ,gen_kwargs={'''files''': files} )] lowercase__ : Optional[Any] = [] for split_name, files in data_files.items(): if isinstance(_snake_case ,_snake_case ): lowercase__ : Tuple = [files] lowercase__ : str = [dl_manager.iter_files(_snake_case ) for file in files] splits.append(datasets.SplitGenerator(name=_snake_case ,gen_kwargs={'''files''': files} ) ) return splits def UpperCAmelCase ( self : int ,_snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowercase__ : List[Any] = self.config.features.arrow_schema.field(_snake_case ).type lowercase__ : Optional[Any] = pa_table.append_column(_snake_case ,pa.array([None] * len(_snake_case ) ,type=_snake_case ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowercase__ : Tuple = table_cast(_snake_case ,self.config.features.arrow_schema ) return pa_table def UpperCAmelCase ( self : str ,_snake_case : int ) -> List[str]: """simple docstring""" for file_idx, file in enumerate(itertools.chain.from_iterable(_snake_case ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(_snake_case ,encoding=self.config.encoding ,errors=self.config.encoding_errors ) as f: lowercase__ : Optional[Any] = json.load(_snake_case ) # We keep only the field we are interested in lowercase__ : str = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(_snake_case ,(list, tuple) ): lowercase__ : Dict = set().union(*[row.keys() for row in dataset] ) lowercase__ : Union[str, Any] = {col: [row.get(_snake_case ) for row in dataset] for col in keys} else: lowercase__ : Union[str, Any] = dataset lowercase__ : Any = pa.Table.from_pydict(_snake_case ) yield file_idx, self._cast_table(_snake_case ) # If the file has one json object per line else: with open(_snake_case ,'''rb''' ) as f: lowercase__ : int = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowercase__ : str = max(self.config.chunksize // 32 ,16 << 10 ) lowercase__ : Union[str, Any] = ( self.config.encoding_errors if self.config.encoding_errors is not None else 'strict' ) while True: lowercase__ : Any = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(_snake_case ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowercase__ : List[Any] = batch.decode(self.config.encoding ,errors=_snake_case ).encode('''utf-8''' ) try: while True: try: lowercase__ : int = paj.read_json( io.BytesIO(_snake_case ) ,read_options=paj.ReadOptions(block_size=_snake_case ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(_snake_case ,pa.ArrowInvalid ) and "straddling" not in str(_snake_case ) or block_size > len(_snake_case ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( f"""Batch of {len(_snake_case )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( _snake_case ,encoding=self.config.encoding ,errors=self.config.encoding_errors ) as f: lowercase__ : Dict = json.load(_snake_case ) except json.JSONDecodeError: logger.error(f"""Failed to read file '{file}' with error {type(_snake_case )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(_snake_case ,_snake_case ): # list is the only sequence type supported in JSON try: lowercase__ : Dict = set().union(*[row.keys() for row in dataset] ) lowercase__ : Optional[int] = {col: [row.get(_snake_case ) for row in dataset] for col in keys} lowercase__ : Optional[Any] = pa.Table.from_pydict(_snake_case ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(f"""Failed to read file '{file}' with error {type(_snake_case )}: {e}""" ) raise ValueError(f"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(_snake_case ) break else: logger.error(f"""Failed to read file '{file}' with error {type(_snake_case )}: {e}""" ) raise ValueError( f"""Not able to read records in the JSON file at {file}. """ f"""You should probably indicate the field of the JSON file containing your records. """ f"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ f"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(_snake_case ) batch_idx += 1
16
def snake_case__ ( SCREAMING_SNAKE_CASE_ : str = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' lowercase__ : str = set() # Replace all the whitespace in our sentence lowercase__ : Tuple = input_str.replace(' ' , '' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(SCREAMING_SNAKE_CASE_ ) == 26 def snake_case__ ( SCREAMING_SNAKE_CASE_ : str = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' lowercase__ : Dict = [False] * 26 for char in input_str: if char.islower(): lowercase__ : List[Any] = True elif char.isupper(): lowercase__ : Optional[Any] = True return all(SCREAMING_SNAKE_CASE_ ) def snake_case__ ( SCREAMING_SNAKE_CASE_ : str = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def snake_case__ ( ): '''simple docstring''' from timeit import timeit lowercase__ : Union[str, Any] = 'from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest' print(timeit('is_pangram()' , setup=SCREAMING_SNAKE_CASE_ ) ) print(timeit('is_pangram_faster()' , setup=SCREAMING_SNAKE_CASE_ ) ) print(timeit('is_pangram_fastest()' , setup=SCREAMING_SNAKE_CASE_ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
214
0
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase): _lowerCamelCase : List[str] = BarthezTokenizer _lowerCamelCase : Optional[int] = BarthezTokenizerFast _lowerCamelCase : Any = True _lowerCamelCase : Any = True def lowercase_ ( self : Any ): """simple docstring""" super().setUp() UpperCamelCase__ = BarthezTokenizerFast.from_pretrained("moussaKam/mbarthez" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname, legacy_format=a_ ) UpperCamelCase__ = tokenizer def lowercase_ ( self : str ): """simple docstring""" UpperCamelCase__ = "<pad>" UpperCamelCase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ), a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ), a_ ) def lowercase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], "<s>" ) self.assertEqual(vocab_keys[1], "<pad>" ) self.assertEqual(vocab_keys[-1], "<mask>" ) self.assertEqual(len(a_ ), 10_1122 ) def lowercase_ ( self : List[str] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size, 10_1122 ) @require_torch def lowercase_ ( self : str ): """simple docstring""" UpperCamelCase__ = ["A long paragraph for summarization.", "Another paragraph for summarization."] UpperCamelCase__ = [0, 57, 3018, 7_0307, 91, 2] UpperCamelCase__ = self.tokenizer( a_, max_length=len(a_ ), padding=a_, truncation=a_, return_tensors="pt" ) self.assertIsInstance(a_, a_ ) self.assertEqual((2, 6), batch.input_ids.shape ) self.assertEqual((2, 6), batch.attention_mask.shape ) UpperCamelCase__ = batch.input_ids.tolist()[0] self.assertListEqual(a_, a_ ) def lowercase_ ( self : int ): """simple docstring""" if not self.test_rust_tokenizer: return UpperCamelCase__ = self.get_tokenizer() UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = "I was born in 92000, and this is falsé." UpperCamelCase__ = tokenizer.tokenize(a_ ) UpperCamelCase__ = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) UpperCamelCase__ = tokenizer.encode(a_, add_special_tokens=a_ ) UpperCamelCase__ = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = tokenizer.encode(a_ ) UpperCamelCase__ = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) @slow def lowercase_ ( self : Dict ): """simple docstring""" UpperCamelCase__ = {"input_ids": [[0, 490, 1_4328, 4507, 354, 47, 4_3669, 95, 25, 7_8117, 2_0215, 1_9779, 190, 22, 400, 4, 3_5343, 8_0310, 603, 86, 2_4937, 105, 3_3438, 9_4762, 196, 3_9642, 7, 15, 1_5933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0534, 87, 25, 66, 3358, 196, 5_5289, 8, 8_2961, 81, 2204, 7_5203, 7, 15, 763, 1_2956, 216, 178, 1_4328, 9595, 1377, 6_9693, 7, 448, 7_1021, 196, 1_8106, 1437, 1_3974, 108, 9083, 4, 4_9315, 7, 39, 86, 1326, 2793, 4_6333, 4, 448, 196, 7_4588, 7, 4_9315, 7, 39, 21, 822, 3_8470, 74, 21, 6_6723, 6_2480, 8, 2_2050, 5, 2]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. UpperCamelCase__ = [ "Le transformeur est un modèle d'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=a_, model_name="moussaKam/mbarthez", revision="c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6", sequences=a_, )
31
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType __lowercase: Optional[List[str]] = None __lowercase: List[Any] = "<" if sys.byteorder == "little" else ">" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image __lowercase: Tuple = [ np.dtype("|b1"), np.dtype("|u1"), np.dtype("<u2"), np.dtype(">u2"), np.dtype("<i2"), np.dtype(">i2"), np.dtype("<u4"), np.dtype(">u4"), np.dtype("<i4"), np.dtype(">i4"), np.dtype("<f4"), np.dtype(">f4"), np.dtype("<f8"), np.dtype(">f8"), ] @dataclass class UpperCAmelCase : _lowerCamelCase : bool = True _lowerCamelCase : Optional[str] = None # Automatically constructed _lowerCamelCase : ClassVar[str] = "PIL.Image.Image" _lowerCamelCase : ClassVar[Any] = pa.struct({'bytes': pa.binary(), 'path': pa.string()}) _lowerCamelCase : str = field(default='Image' , init=SCREAMING_SNAKE_CASE__ , repr=SCREAMING_SNAKE_CASE__) def __call__( self : Union[str, Any] ): """simple docstring""" return self.pa_type def lowercase_ ( self : Optional[Any], a_ : Union[str, bytes, dict, np.ndarray, "PIL.Image.Image"] ): """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if isinstance(a_, a_ ): UpperCamelCase__ = np.array(a_ ) if isinstance(a_, a_ ): return {"path": value, "bytes": None} elif isinstance(a_, a_ ): return {"path": None, "bytes": value} elif isinstance(a_, np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(a_ ) elif isinstance(a_, PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(a_ ) elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( f'An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def lowercase_ ( self : Dict, a_ : dict, a_ : Dict=None ): """simple docstring""" if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Image(decode=True) instead." ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support decoding images, please install 'Pillow'." ) if token_per_repo_id is None: UpperCamelCase__ = {} UpperCamelCase__ , UpperCamelCase__ = value["path"], value["bytes"] if bytes_ is None: if path is None: raise ValueError(f'An image should have one of \'path\' or \'bytes\' but both are None in {value}.' ) else: if is_local_path(a_ ): UpperCamelCase__ = PIL.Image.open(a_ ) else: UpperCamelCase__ = path.split("::" )[-1] try: UpperCamelCase__ = string_to_dict(a_, config.HUB_DATASETS_URL )["repo_id"] UpperCamelCase__ = token_per_repo_id.get(a_ ) except ValueError: UpperCamelCase__ = None with xopen(a_, "rb", use_auth_token=a_ ) as f: UpperCamelCase__ = BytesIO(f.read() ) UpperCamelCase__ = PIL.Image.open(bytes_ ) else: UpperCamelCase__ = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def lowercase_ ( self : List[str] ): """simple docstring""" from .features import Value return ( self if self.decode else { "bytes": Value("binary" ), "path": Value("string" ), } ) def lowercase_ ( self : List[Any], a_ : Union[pa.StringArray, pa.StructArray, pa.ListArray] ): """simple docstring""" if pa.types.is_string(storage.type ): UpperCamelCase__ = pa.array([None] * len(a_ ), type=pa.binary() ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, storage], ["bytes", "path"], mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCamelCase__ = pa.array([None] * len(a_ ), type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays([storage, path_array], ["bytes", "path"], mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: UpperCamelCase__ = storage.field("bytes" ) else: UpperCamelCase__ = pa.array([None] * len(a_ ), type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: UpperCamelCase__ = storage.field("path" ) else: UpperCamelCase__ = pa.array([None] * len(a_ ), type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, path_array], ["bytes", "path"], mask=storage.is_null() ) elif pa.types.is_list(storage.type ): UpperCamelCase__ = pa.array( [encode_np_array(np.array(a_ ) )["bytes"] if arr is not None else None for arr in storage.to_pylist()], type=pa.binary(), ) UpperCamelCase__ = pa.array([None] * len(a_ ), type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays( [bytes_array, path_array], ["bytes", "path"], mask=bytes_array.is_null() ) return array_cast(a_, self.pa_type ) def lowercase_ ( self : str, a_ : pa.StructArray ): """simple docstring""" @no_op_if_value_is_null def path_to_bytes(a_ : Dict ): with xopen(a_, "rb" ) as f: UpperCamelCase__ = f.read() return bytes_ UpperCamelCase__ = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ], type=pa.binary(), ) UpperCamelCase__ = pa.array( [os.path.basename(a_ ) if path is not None else None for path in storage.field("path" ).to_pylist()], type=pa.string(), ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, path_array], ["bytes", "path"], mask=bytes_array.is_null() ) return array_cast(a_, self.pa_type ) def SCREAMING_SNAKE_CASE__( ) -> List[str]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() UpperCamelCase__ = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def SCREAMING_SNAKE_CASE__( _UpperCamelCase : "PIL.Image.Image" ) -> bytes: '''simple docstring''' UpperCamelCase__ = BytesIO() if image.format in list_image_compression_formats(): UpperCamelCase__ = image.format else: UpperCamelCase__ = "PNG" if image.mode in ["1", "L", "LA", "RGB", "RGBA"] else "TIFF" image.save(_UpperCamelCase , format=_UpperCamelCase ) return buffer.getvalue() def SCREAMING_SNAKE_CASE__( _UpperCamelCase : "PIL.Image.Image" ) -> dict: '''simple docstring''' if hasattr(_UpperCamelCase , "filename" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(_UpperCamelCase )} def SCREAMING_SNAKE_CASE__( _UpperCamelCase : np.ndarray ) -> dict: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) UpperCamelCase__ = array.dtype UpperCamelCase__ = dtype.byteorder if dtype.byteorder != "=" else _NATIVE_BYTEORDER UpperCamelCase__ = dtype.kind UpperCamelCase__ = dtype.itemsize UpperCamelCase__ = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: UpperCamelCase__ = np.dtype("|u1" ) if dtype_kind not in ["u", "i"]: raise TypeError( F'Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.' ) if dtype is not dest_dtype: warnings.warn(F'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: UpperCamelCase__ = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: UpperCamelCase__ = dtype_byteorder + dtype_kind + str(_UpperCamelCase ) UpperCamelCase__ = np.dtype(_UpperCamelCase ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}' ) UpperCamelCase__ = PIL.Image.fromarray(array.astype(_UpperCamelCase ) ) return {"path": None, "bytes": image_to_bytes(_UpperCamelCase )} def SCREAMING_SNAKE_CASE__( _UpperCamelCase : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ) -> List[dict]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if objs: UpperCamelCase__ , UpperCamelCase__ = first_non_null_value(_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(_UpperCamelCase , np.ndarray ): UpperCamelCase__ = no_op_if_value_is_null(_UpperCamelCase ) return [obj_to_image_dict_func(_UpperCamelCase ) for obj in objs] elif isinstance(_UpperCamelCase , PIL.Image.Image ): UpperCamelCase__ = no_op_if_value_is_null(_UpperCamelCase ) return [obj_to_image_dict_func(_UpperCamelCase ) for obj in objs] else: return objs else: return objs
31
1
class _SCREAMING_SNAKE_CASE : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )-> Tuple: lowerCamelCase_ =None lowerCamelCase_ =None lowerCamelCase_ =graph self._normalize_graph(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCamelCase_ =len(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =None def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )-> int: if sources is int: lowerCamelCase_ =[sources] if sinks is int: lowerCamelCase_ =[sinks] if len(_SCREAMING_SNAKE_CASE ) == 0 or len(_SCREAMING_SNAKE_CASE ) == 0: return lowerCamelCase_ =sources[0] lowerCamelCase_ =sinks[0] # make fake vertex if there are more # than one source or sink if len(_SCREAMING_SNAKE_CASE ) > 1 or len(_SCREAMING_SNAKE_CASE ) > 1: lowerCamelCase_ =0 for i in sources: max_input_flow += sum(self.graph[i] ) lowerCamelCase_ =len(self.graph ) + 1 for room in self.graph: room.insert(0 , 0 ) self.graph.insert(0 , [0] * size ) for i in sources: lowerCamelCase_ =max_input_flow lowerCamelCase_ =0 lowerCamelCase_ =len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: lowerCamelCase_ =max_input_flow lowerCamelCase_ =size - 1 def _snake_case ( self )-> List[str]: if self.maximum_flow_algorithm is None: raise Exception("""You need to set maximum flow algorithm before.""" ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> int: lowerCamelCase_ =algorithm(self ) class _SCREAMING_SNAKE_CASE : def __init__( self , _SCREAMING_SNAKE_CASE )-> List[str]: lowerCamelCase_ =flow_network lowerCamelCase_ =flow_network.verticesCount lowerCamelCase_ =flow_network.sourceIndex lowerCamelCase_ =flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that lowerCamelCase_ =flow_network.graph lowerCamelCase_ =False def _snake_case ( self )-> List[str]: if not self.executed: self._algorithm() lowerCamelCase_ =True def _snake_case ( self )-> Optional[Any]: pass class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def __init__( self , _SCREAMING_SNAKE_CASE )-> Optional[Any]: super().__init__(_SCREAMING_SNAKE_CASE ) # use this to save your result lowerCamelCase_ =-1 def _snake_case ( self )-> Tuple: if not self.executed: raise Exception("""You should execute algorithm before using its result!""" ) return self.maximum_flow class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def __init__( self , _SCREAMING_SNAKE_CASE )-> str: super().__init__(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =[[0] * self.verticies_count for i in range(self.verticies_count )] lowerCamelCase_ =[0] * self.verticies_count lowerCamelCase_ =[0] * self.verticies_count def _snake_case ( self )-> Tuple: lowerCamelCase_ =self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule lowerCamelCase_ =[ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list lowerCamelCase_ =0 while i < len(_SCREAMING_SNAKE_CASE ): lowerCamelCase_ =vertices_list[i] lowerCamelCase_ =self.heights[vertex_index] self.process_vertex(_SCREAMING_SNAKE_CASE ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(_SCREAMING_SNAKE_CASE ) ) lowerCamelCase_ =0 else: i += 1 lowerCamelCase_ =sum(self.preflow[self.source_index] ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Optional[Any]: while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.relabel(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )-> Optional[int]: lowerCamelCase_ =min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Optional[int]: lowerCamelCase_ =None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): lowerCamelCase_ =self.heights[to_index] if min_height is not None: lowerCamelCase_ =min_height + 1 if __name__ == "__main__": __A : int = [0] __A : Union[str, Any] = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] __A : Optional[Any] = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network __A : str = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate __A : Dict = flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
154
import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() __A : Tuple = logging.get_logger(__name__) def __UpperCamelCase ( _A : str , _A : str , _A : str ) ->int: """simple docstring""" lowerCamelCase_ =UniSpeechSatForSequenceClassification.from_pretrained(_A , config=_A ) lowerCamelCase_ =downstream_dict["""projector.weight"""] lowerCamelCase_ =downstream_dict["""projector.bias"""] lowerCamelCase_ =downstream_dict["""model.post_net.linear.weight"""] lowerCamelCase_ =downstream_dict["""model.post_net.linear.bias"""] return model def __UpperCamelCase ( _A : Optional[int] , _A : str , _A : Any ) ->Optional[int]: """simple docstring""" lowerCamelCase_ =UniSpeechSatForAudioFrameClassification.from_pretrained(_A , config=_A ) lowerCamelCase_ =downstream_dict["""model.linear.weight"""] lowerCamelCase_ =downstream_dict["""model.linear.bias"""] return model def __UpperCamelCase ( _A : Optional[Any] , _A : Optional[Any] , _A : Optional[Any] ) ->List[Any]: """simple docstring""" lowerCamelCase_ =UniSpeechSatForXVector.from_pretrained(_A , config=_A ) lowerCamelCase_ =downstream_dict["""connector.weight"""] lowerCamelCase_ =downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): lowerCamelCase_ =downstream_dict[ f'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] lowerCamelCase_ =downstream_dict[f'model.framelevel_feature_extractor.module.{i}.kernel.bias'] lowerCamelCase_ =downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] lowerCamelCase_ =downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] lowerCamelCase_ =downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] lowerCamelCase_ =downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] lowerCamelCase_ =downstream_dict["""objective.W"""] return model @torch.no_grad() def __UpperCamelCase ( _A : Any , _A : Optional[Any] , _A : Union[str, Any] , _A : str ) ->Union[str, Any]: """simple docstring""" lowerCamelCase_ =torch.load(_A , map_location="""cpu""" ) lowerCamelCase_ =checkpoint["""Downstream"""] lowerCamelCase_ =UniSpeechSatConfig.from_pretrained(_A ) lowerCamelCase_ =WavaVecaFeatureExtractor.from_pretrained( _A , return_attention_mask=_A , do_normalize=_A ) lowerCamelCase_ =hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): lowerCamelCase_ =convert_classification(_A , _A , _A ) elif arch.endswith("""ForAudioFrameClassification""" ): lowerCamelCase_ =convert_diarization(_A , _A , _A ) elif arch.endswith("""ForXVector""" ): lowerCamelCase_ =convert_xvector(_A , _A , _A ) else: raise NotImplementedError(f'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: lowerCamelCase_ =checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(_A ) hf_model.save_pretrained(_A ) if __name__ == "__main__": __A : Optional[int] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') __A : int = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
154
1
"""simple docstring""" from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging SCREAMING_SNAKE_CASE__:Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Dict = { """google/umt5-small""": """https://huggingface.co/google/umt5-small/resolve/main/config.json""", # See all umt5 models at https://huggingface.co/models?filter=umt5 } class snake_case__ ( snake_case_ ): _snake_case : int = """umt5""" _snake_case : str = ["""past_key_values"""] def __init__( self , lowerCamelCase=250112 , lowerCamelCase=512 , lowerCamelCase=64 , lowerCamelCase=1024 , lowerCamelCase=8 , lowerCamelCase=None , lowerCamelCase=6 , lowerCamelCase=32 , lowerCamelCase=128 , lowerCamelCase=0.1 , lowerCamelCase=1E-6 , lowerCamelCase=1.0 , lowerCamelCase="gated-gelu" , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="T5Tokenizer" , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=0 , **lowerCamelCase , ): super().__init__( is_encoder_decoder=lowerCamelCase , tokenizer_class=lowerCamelCase , tie_word_embeddings=lowerCamelCase , pad_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , decoder_start_token_id=lowerCamelCase , **lowerCamelCase , ) __a = vocab_size __a = d_model __a = d_kv __a = d_ff __a = num_layers __a = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __a = num_heads __a = relative_attention_num_buckets __a = relative_attention_max_distance __a = dropout_rate __a = layer_norm_epsilon __a = initializer_factor __a = feed_forward_proj __a = use_cache __a = self.feed_forward_proj.split("-" ) __a = act_info[-1] __a = act_info[0] == "gated" if len(lowerCamelCase ) > 1 and act_info[0] != "gated" or len(lowerCamelCase ) > 2: raise ValueError( F"`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer." "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "'gated-gelu' or 'relu'" ) if feed_forward_proj == "gated-gelu": __a = "gelu_new" @property def a__ ( self ): return self.d_model @property def a__ ( self ): return self.num_heads @property def a__ ( self ): return self.num_layers class snake_case__ ( snake_case_ ): @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def a__ ( self ): __a = { "input_ids": {0: "batch", 1: "encoder_sequence"}, "attention_mask": {0: "batch", 1: "encoder_sequence"}, } if self.use_past: __a = "past_encoder_sequence + sequence" __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def a__ ( self ): return 13 @property def a__ ( self ): return 5E-4
268
"""simple docstring""" from ..utils import DummyObject, requires_backends class snake_case__ ( metaclass=snake_case_ ): _snake_case : Union[str, Any] = ["""onnx"""] def __init__( self , *lowerCamelCase , **lowerCamelCase ): requires_backends(self , ["onnx"] ) @classmethod def a__ ( cls , *lowerCamelCase , **lowerCamelCase ): requires_backends(cls , ["onnx"] ) @classmethod def a__ ( cls , *lowerCamelCase , **lowerCamelCase ): requires_backends(cls , ["onnx"] )
268
1
from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging SCREAMING_SNAKE_CASE :Optional[Any] = logging.get_logger(__name__) class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = ["pixel_values"] def __init__( self : Dict ,A : bool = True ,A : Dict[str, int] = None ,A : PILImageResampling = PILImageResampling.BICUBIC ,A : bool = True ,A : Dict[str, int] = None ,A : bool = True ,A : Union[int, float] = 1 / 2_55 ,A : bool = True ,A : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_MEAN ,A : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_STD ,**A : Dict ,): super().__init__(**A ) __A = size if size is not None else {"shortest_edge": 2_24} __A = get_size_dict(A ,default_to_square=A ) __A = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} __A = get_size_dict(A ,param_name="crop_size" ) __A = do_resize __A = size __A = resample __A = do_center_crop __A = crop_size __A = do_rescale __A = rescale_factor __A = do_normalize __A = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __A = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCamelCase_ ( self : Union[str, Any] ,A : np.ndarray ,A : Dict[str, int] ,A : PILImageResampling = PILImageResampling.BICUBIC ,A : Optional[Union[str, ChannelDimension]] = None ,**A : Optional[Any] ,): __A = get_size_dict(A ,default_to_square=A ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __A = int((2_56 / 2_24) * size["shortest_edge"] ) __A = get_resize_output_image_size(A ,size=A ,default_to_square=A ) __A = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f'''Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}''' ) return resize( A ,size=(size_dict["height"], size_dict["width"]) ,resample=A ,data_format=A ,**A ) def UpperCamelCase_ ( self : Union[str, Any] ,A : np.ndarray ,A : Dict[str, int] ,A : Optional[Union[str, ChannelDimension]] = None ,**A : Tuple ,): __A = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(f'''Size dict must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return center_crop(A ,size=(size["height"], size["width"]) ,data_format=A ,**A ) def UpperCamelCase_ ( self : Any ,A : np.ndarray ,A : Union[int, float] ,A : Optional[Union[str, ChannelDimension]] = None ,**A : str ,): return rescale(A ,scale=A ,data_format=A ,**A ) def UpperCamelCase_ ( self : List[Any] ,A : np.ndarray ,A : Union[float, List[float]] ,A : Union[float, List[float]] ,A : Optional[Union[str, ChannelDimension]] = None ,**A : str ,): return normalize(A ,mean=A ,std=A ,data_format=A ,**A ) def UpperCamelCase_ ( self : int ,A : ImageInput ,A : Optional[bool] = None ,A : Optional[Dict[str, int]] = None ,A : PILImageResampling = None ,A : Optional[bool] = None ,A : Optional[Dict[str, int]] = None ,A : Optional[bool] = None ,A : Optional[float] = None ,A : Optional[bool] = None ,A : Optional[Union[float, Iterable[float]]] = None ,A : Optional[Union[float, Iterable[float]]] = None ,A : Optional[TensorType] = None ,A : ChannelDimension = ChannelDimension.FIRST ,**A : str ,): __A = do_resize if do_resize is not None else self.do_resize __A = resample if resample is not None else self.resample __A = do_center_crop if do_center_crop is not None else self.do_center_crop __A = do_rescale if do_rescale is not None else self.do_rescale __A = rescale_factor if rescale_factor is not None else self.rescale_factor __A = do_normalize if do_normalize is not None else self.do_normalize __A = image_mean if image_mean is not None else self.image_mean __A = image_std if image_std is not None else self.image_std __A = size if size is not None else self.size __A = get_size_dict(A ,default_to_square=A ) __A = crop_size if crop_size is not None else self.crop_size __A = get_size_dict(A ,param_name="crop_size" ) __A = make_list_of_images(A ) if not valid_images(A ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __A = [to_numpy_array(A ) for image in images] if do_resize: __A = [self.resize(A ,A ,A ) for image in images] if do_center_crop: __A = [self.center_crop(A ,A ) for image in images] if do_rescale: __A = [self.rescale(A ,A ) for image in images] if do_normalize: __A = [self.normalize(A ,A ,A ) for image in images] __A = [to_channel_dimension_format(A ,A ) for image in images] __A = {"pixel_values": images} return BatchFeature(data=A ,tensor_type=A )
15
import math def A ( _lowercase ): return math.sqrt(_lowercase ) * math.sqrt(_lowercase ) == num def A ( _lowercase ): SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Tuple = n while left <= right: SCREAMING_SNAKE_CASE : Union[str, Any] = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: SCREAMING_SNAKE_CASE : Optional[Any] = mid - 1 else: SCREAMING_SNAKE_CASE : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
182
0
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow __UpperCAmelCase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ "text-classification", "language-modeling", "summarization", "token-classification", "question-answering", ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) __UpperCAmelCase = logging.getLogger() def A__ ( ): SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument('''-f''' ) SCREAMING_SNAKE_CASE_ = parser.parse_args() return args.f def A__ ( __lowerCamelCase, __lowerCamelCase="eval" ): SCREAMING_SNAKE_CASE_ = os.path.join(__lowerCAmelCase, F'''{split}_results.json''' ) if os.path.exists(__lowerCAmelCase ): with open(__lowerCAmelCase, '''r''' ) as f: return json.load(__lowerCAmelCase ) raise ValueError(F'''can\'t find {path}''' ) __UpperCAmelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class UpperCamelCase__ ( A__ ): """simple docstring""" def _UpperCamelCase ( self ) -> Dict: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_flax_glue.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) @slow def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_clm_flax.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertLess(result['''eval_perplexity'''] , 100 ) @slow def _UpperCamelCase ( self ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_summarization_flax.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def _UpperCamelCase ( self ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_mlm_flax.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_ta_mlm_flax.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.42 ) @slow def _UpperCamelCase ( self ) -> Union[str, Any]: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu SCREAMING_SNAKE_CASE_ = 7 if get_gpu_count() > 1 else 2 SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_flax_ner.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE_ = F''' run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 '''.split() with patch.object(__snake_case , '''argv''' , __snake_case ): run_qa.main() SCREAMING_SNAKE_CASE_ = get_results(__snake_case ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
353
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": __UpperCAmelCase = input("Enter image url: ").strip() print(F"""Downloading image from {url} ...""") __UpperCAmelCase = BeautifulSoup(requests.get(url).content, "html.parser") # The image URL is in the content field of the first meta tag with property og:image __UpperCAmelCase = soup.find("meta", {"property": "og:image"})["content"] __UpperCAmelCase = requests.get(image_url).content __UpperCAmelCase = F"""{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg""" with open(file_name, "wb") as fp: fp.write(image_data) print(F"""Done. Image saved to disk as {file_name}.""")
257
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Tuple = { """google/bit-50""": """https://huggingface.co/google/bit-50/resolve/main/config.json""", } class lowercase ( lowercase_ , lowercase_ ): __SCREAMING_SNAKE_CASE : List[str] = '''bit''' __SCREAMING_SNAKE_CASE : Dict = ['''preactivation''', '''bottleneck'''] __SCREAMING_SNAKE_CASE : Any = ['''SAME''', '''VALID'''] def __init__( self , snake_case=3 , snake_case=64 , snake_case=[256, 512, 1024, 2048] , snake_case=[3, 4, 6, 3] , snake_case="preactivation" , snake_case="relu" , snake_case=None , snake_case=32 , snake_case=0.0 , snake_case=False , snake_case=32 , snake_case=1 , snake_case=None , snake_case=None , **snake_case , ): super().__init__(**snake_case ) if layer_type not in self.layer_types: raise ValueError(F'''layer_type={layer_type} is not one of {",".join(self.layer_types )}''' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case_ = global_padding.upper() else: raise ValueError(F'''Padding strategy {global_padding} not supported''' ) snake_case_ = num_channels snake_case_ = embedding_size snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = layer_type snake_case_ = hidden_act snake_case_ = global_padding snake_case_ = num_groups snake_case_ = drop_path_rate snake_case_ = embedding_dynamic_padding snake_case_ = output_stride snake_case_ = width_factor snake_case_ = ['stem'] + [F'''stage{idx}''' for idx in range(1 , len(snake_case ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=snake_case , out_indices=snake_case , stage_names=self.stage_names )
285
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: snake_case_ = mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else: snake_case_ = max( mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , j - wt[i - 1] ) + val[i - 1] , ) snake_case_ = val return f[i][j] def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: snake_case_ = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: snake_case_ = dp[i - 1][w_] return dp[n][w_], dp def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if not (isinstance(UpperCamelCase__ , (list, tuple) ) and isinstance(UpperCamelCase__ , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) snake_case_ = len(UpperCamelCase__ ) if num_items != len(UpperCamelCase__ ): snake_case_ = ( 'The number of weights must be the same as the number of values.\n' F'''But got {num_items} weights and {len(UpperCamelCase__ )} values''' ) raise ValueError(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): if not isinstance(wt[i] , UpperCamelCase__ ): snake_case_ = ( 'All weights must be integers but got weight of ' F'''type {type(wt[i] )} at index {i}''' ) raise TypeError(UpperCamelCase__ ) snake_case_ , snake_case_ = knapsack(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = set() _construct_solution(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return optimal_val, example_optional_set def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , UpperCamelCase__ , UpperCamelCase__ ) else: optimal_set.add(UpperCamelCase__ ) _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , j - wt[i - 1] , UpperCamelCase__ ) if __name__ == "__main__": _UpperCAmelCase : Tuple = [3, 2, 4, 4] _UpperCAmelCase : Optional[Any] = [4, 3, 2, 3] _UpperCAmelCase : List[str] = 4 _UpperCAmelCase : str = 6 _UpperCAmelCase : Tuple = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _UpperCAmelCase , _UpperCAmelCase : List[Any] = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _UpperCAmelCase , _UpperCAmelCase : Any = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("""optimal_value = """, optimal_solution) print("""An optimal subset corresponding to the optimal value""", optimal_subset)
285
1
"""simple docstring""" import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class __snake_case : def __init__( self : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : Union[str, Any]=14 , _UpperCAmelCase : Union[str, Any]=7 , _UpperCAmelCase : Optional[Any]=True , _UpperCAmelCase : Optional[int]=True , _UpperCAmelCase : List[Any]=False , _UpperCAmelCase : Optional[int]=True , _UpperCAmelCase : Optional[int]=99 , _UpperCAmelCase : List[str]=32 , _UpperCAmelCase : List[Any]=4 , _UpperCAmelCase : Optional[int]=4 , _UpperCAmelCase : Tuple=4 , _UpperCAmelCase : List[str]=37 , _UpperCAmelCase : int="gelu" , _UpperCAmelCase : List[Any]=0.1 , _UpperCAmelCase : Dict=0.1 , _UpperCAmelCase : str=512 , _UpperCAmelCase : Optional[int]=0.02 , ) -> Any: '''simple docstring''' _lowerCAmelCase : int = parent _lowerCAmelCase : int = batch_size _lowerCAmelCase : Optional[Any] = seq_length _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Union[str, Any] = use_input_mask _lowerCAmelCase : Dict = use_token_type_ids _lowerCAmelCase : Optional[int] = use_labels _lowerCAmelCase : Optional[Any] = vocab_size _lowerCAmelCase : List[Any] = hidden_size _lowerCAmelCase : Any = rotary_dim _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : Optional[Any] = num_attention_heads _lowerCAmelCase : Optional[Any] = intermediate_size _lowerCAmelCase : Tuple = hidden_act _lowerCAmelCase : Dict = hidden_dropout_prob _lowerCAmelCase : int = attention_probs_dropout_prob _lowerCAmelCase : Dict = max_position_embeddings _lowerCAmelCase : Tuple = initializer_range _lowerCAmelCase : Dict = None _lowerCAmelCase : int = vocab_size - 1 _lowerCAmelCase : List[Any] = vocab_size - 1 _lowerCAmelCase : Dict = vocab_size - 1 def SCREAMING_SNAKE_CASE ( self : str ) -> Any: '''simple docstring''' _lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : List[str] = None if self.use_input_mask: _lowerCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase : Union[str, Any] = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_UpperCAmelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Dict: '''simple docstring''' _lowerCAmelCase : str = self.prepare_config_and_inputs() _lowerCAmelCase : Optional[Any] = config_and_inputs _lowerCAmelCase : Tuple = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self : Optional[Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : str , _UpperCAmelCase : List[Any] ) -> Any: '''simple docstring''' _lowerCAmelCase : str = 20 _lowerCAmelCase : Tuple = model_class_name(_UpperCAmelCase ) _lowerCAmelCase : str = model.init_cache(input_ids.shape[0] , _UpperCAmelCase ) _lowerCAmelCase : Optional[int] = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _lowerCAmelCase : Union[str, Any] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) _lowerCAmelCase : Dict = model( input_ids[:, :-1] , attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , position_ids=_UpperCAmelCase , ) _lowerCAmelCase : int = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="""i4""" ) _lowerCAmelCase : List[Any] = model( input_ids[:, -1:] , attention_mask=_UpperCAmelCase , past_key_values=outputs_cache.past_key_values , position_ids=_UpperCAmelCase , ) _lowerCAmelCase : str = model(_UpperCAmelCase ) _lowerCAmelCase : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"Max diff is {diff}" ) def SCREAMING_SNAKE_CASE ( self : int , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[int] ) -> int: '''simple docstring''' _lowerCAmelCase : Optional[Any] = 20 _lowerCAmelCase : List[str] = model_class_name(_UpperCAmelCase ) _lowerCAmelCase : List[str] = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) _lowerCAmelCase : Any = model.init_cache(input_ids.shape[0] , _UpperCAmelCase ) _lowerCAmelCase : Tuple = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) _lowerCAmelCase : Any = model( input_ids[:, :-1] , attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , position_ids=_UpperCAmelCase , ) _lowerCAmelCase : List[Any] = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="""i4""" ) _lowerCAmelCase : Optional[int] = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_UpperCAmelCase , position_ids=_UpperCAmelCase , ) _lowerCAmelCase : Dict = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase ) _lowerCAmelCase : Any = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"Max diff is {diff}" ) @require_flax class __snake_case (_a , _a , unittest.TestCase ): lowerCAmelCase__ = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () lowerCAmelCase__ = (FlaxGPTJForCausalLM,) if is_flax_available() else () def SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: '''simple docstring''' _lowerCAmelCase : Dict = FlaxGPTJModelTester(self ) def SCREAMING_SNAKE_CASE ( self : int ) -> Union[str, Any]: '''simple docstring''' for model_class_name in self.all_model_classes: _lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) @tooslow def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: '''simple docstring''' _lowerCAmelCase : str = GPTaTokenizer.from_pretrained("""gpt2""" , pad_token="""<|endoftext|>""" , padding_side="""left""" ) _lowerCAmelCase : Optional[Any] = tokenizer(["""Hello this is a long string""", """Hey"""] , return_tensors="""np""" , padding=_UpperCAmelCase , truncation=_UpperCAmelCase ) _lowerCAmelCase : int = FlaxGPTJForCausalLM.from_pretrained("""EleutherAI/gpt-j-6B""" ) _lowerCAmelCase : str = False _lowerCAmelCase : Union[str, Any] = model.config.eos_token_id _lowerCAmelCase : Tuple = jax.jit(model.generate ) _lowerCAmelCase : List[str] = jit_generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , pad_token_id=tokenizer.pad_token_id ).sequences _lowerCAmelCase : Union[str, Any] = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) _lowerCAmelCase : str = [ """Hello this is a long string of text.\n\nI'm trying to get the text of the""", """Hey, I'm a little late to the party. I'm going to""", ] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @is_pt_flax_cross_test def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: '''simple docstring''' _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs _lowerCAmelCase : Tuple = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Dict = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class _lowerCAmelCase : List[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowerCAmelCase : Optional[Any] = getattr(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : List[Any] = pt_inputs["""input_ids"""].shape _lowerCAmelCase : List[str] = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_UpperCAmelCase ): _lowerCAmelCase : List[Any] = 0 _lowerCAmelCase : str = 1 _lowerCAmelCase : Optional[int] = 0 _lowerCAmelCase : List[str] = 1 _lowerCAmelCase : Union[str, Any] = pt_model_class(_UpperCAmelCase ).eval() _lowerCAmelCase : Optional[int] = model_class(_UpperCAmelCase , dtype=jnp.floataa ) _lowerCAmelCase : Any = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _UpperCAmelCase ) _lowerCAmelCase : Union[str, Any] = fx_state with torch.no_grad(): _lowerCAmelCase : Any = pt_model(**_UpperCAmelCase ).to_tuple() _lowerCAmelCase : str = fx_model(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_UpperCAmelCase ) _lowerCAmelCase : Dict = model_class.from_pretrained(_UpperCAmelCase , from_pt=_UpperCAmelCase ) _lowerCAmelCase : Union[str, Any] = fx_model_loaded(**_UpperCAmelCase ).to_tuple() self.assertEqual( len(_UpperCAmelCase ) , len(_UpperCAmelCase ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output_loaded, pt_output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> str: '''simple docstring''' _lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs _lowerCAmelCase : List[str] = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : str = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class _lowerCAmelCase : List[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowerCAmelCase : str = getattr(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Tuple = pt_model_class(_UpperCAmelCase ).eval() _lowerCAmelCase : Union[str, Any] = model_class(_UpperCAmelCase , dtype=jnp.floataa ) _lowerCAmelCase : Any = load_flax_weights_in_pytorch_model(_UpperCAmelCase , fx_model.params ) _lowerCAmelCase : Optional[Any] = pt_inputs["""input_ids"""].shape _lowerCAmelCase : int = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_UpperCAmelCase ): _lowerCAmelCase : Tuple = 0 _lowerCAmelCase : Tuple = 1 _lowerCAmelCase : Dict = 0 _lowerCAmelCase : List[str] = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): _lowerCAmelCase : List[Any] = pt_model(**_UpperCAmelCase ).to_tuple() _lowerCAmelCase : List[Any] = fx_model(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_UpperCAmelCase ) _lowerCAmelCase : Any = pt_model_class.from_pretrained(_UpperCAmelCase , from_flax=_UpperCAmelCase ) with torch.no_grad(): _lowerCAmelCase : str = pt_model_loaded(**_UpperCAmelCase ).to_tuple() self.assertEqual( len(_UpperCAmelCase ) , len(_UpperCAmelCase ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def SCREAMING_SNAKE_CASE ( self : List[str] ) -> int: '''simple docstring''' for model_class_name in self.all_model_classes: _lowerCAmelCase : Optional[Any] = model_class_name.from_pretrained("""EleutherAI/gpt-j-6B""" ) _lowerCAmelCase : List[str] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_UpperCAmelCase )
362
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _lowerCamelCase : Any = logging.get_logger(__name__) _lowerCamelCase : Optional[int] = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _lowerCamelCase : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __snake_case : lowerCAmelCase__ = field( default=_a , metadata={"help": "Model type selected in the list: " + ", ".join(_a )} ) lowerCAmelCase__ = field( default=_a , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) lowerCAmelCase__ = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowerCAmelCase__ = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) lowerCAmelCase__ = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) lowerCAmelCase__ = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) lowerCAmelCase__ = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"} ) lowerCAmelCase__ = field( default=_a , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) lowerCAmelCase__ = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) lowerCAmelCase__ = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) lowerCAmelCase__ = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) lowerCAmelCase__ = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class __snake_case (_a ): lowerCAmelCase__ = "train" lowerCAmelCase__ = "dev" class __snake_case (_a ): lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 def __init__( self : Tuple , _UpperCAmelCase : SquadDataTrainingArguments , _UpperCAmelCase : PreTrainedTokenizer , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : Union[str, Split] = Split.train , _UpperCAmelCase : Optional[bool] = False , _UpperCAmelCase : Optional[str] = None , _UpperCAmelCase : Optional[str] = "pt" , ) -> Optional[Any]: '''simple docstring''' _lowerCAmelCase : Optional[Any] = args _lowerCAmelCase : Union[str, Any] = is_language_sensitive _lowerCAmelCase : Dict = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_UpperCAmelCase , _UpperCAmelCase ): try: _lowerCAmelCase : List[str] = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) _lowerCAmelCase : List[str] = mode # Load data features from cache or dataset file _lowerCAmelCase : List[Any] = """v2""" if args.version_2_with_negative else """v1""" _lowerCAmelCase : Any = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _lowerCAmelCase : Union[str, Any] = cached_features_file + """.lock""" with FileLock(_UpperCAmelCase ): if os.path.exists(_UpperCAmelCase ) and not args.overwrite_cache: _lowerCAmelCase : Dict = time.time() _lowerCAmelCase : Any = torch.load(_UpperCAmelCase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. _lowerCAmelCase : List[str] = self.old_features["""features"""] _lowerCAmelCase : List[Any] = self.old_features.get("""dataset""" , _UpperCAmelCase ) _lowerCAmelCase : int = self.old_features.get("""examples""" , _UpperCAmelCase ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" """ future run""" ) else: if mode == Split.dev: _lowerCAmelCase : Optional[Any] = self.processor.get_dev_examples(args.data_dir ) else: _lowerCAmelCase : Optional[Any] = self.processor.get_train_examples(args.data_dir ) _lowerCAmelCase , _lowerCAmelCase : List[str] = squad_convert_examples_to_features( examples=self.examples , tokenizer=_UpperCAmelCase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_UpperCAmelCase , ) _lowerCAmelCase : List[str] = time.time() torch.save( {"""features""": self.features, """dataset""": self.dataset, """examples""": self.examples} , _UpperCAmelCase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' return len(self.features ) def __getitem__( self : Any , _UpperCAmelCase : Dict ) -> Dict[str, torch.Tensor]: '''simple docstring''' _lowerCAmelCase : str = self.features[i] _lowerCAmelCase : Tuple = torch.tensor(feature.input_ids , dtype=torch.long ) _lowerCAmelCase : Tuple = torch.tensor(feature.attention_mask , dtype=torch.long ) _lowerCAmelCase : Union[str, Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) _lowerCAmelCase : int = torch.tensor(feature.cls_index , dtype=torch.long ) _lowerCAmelCase : Optional[int] = torch.tensor(feature.p_mask , dtype=torch.float ) _lowerCAmelCase : Union[str, Any] = torch.tensor(feature.is_impossible , dtype=torch.float ) _lowerCAmelCase : Tuple = { """input_ids""": input_ids, """attention_mask""": attention_mask, """token_type_ids""": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"""cls_index""": cls_index, """p_mask""": p_mask} ) if self.args.version_2_with_negative: inputs.update({"""is_impossible""": is_impossible} ) if self.is_language_sensitive: inputs.update({"""langs""": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: _lowerCAmelCase : Union[str, Any] = torch.tensor(feature.start_position , dtype=torch.long ) _lowerCAmelCase : Union[str, Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"""start_positions""": start_positions, """end_positions""": end_positions} ) return inputs
159
0
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowerCamelCase : Optional[Any] = logging.get_logger(__name__) _lowerCamelCase : List[Any] = {"""vocab_file""": """vocab.txt"""} _lowerCamelCase : List[Any] = { """vocab_file""": { """openbmb/cpm-ant-10b""": """https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt""", }, } _lowerCamelCase : List[str] = { """openbmb/cpm-ant-10b""": 1024, } def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Any: """simple docstring""" A__ = collections.OrderedDict() with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as reader: A__ = reader.readlines() for index, token in enumerate(lowercase_ ): A__ = token.rstrip('''\n''' ) A__ = index return vocab class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str]="<unk>" , UpperCAmelCase__ : Union[str, Any]=200) ->List[str]: '''simple docstring''' A__ = vocab A__ = unk_token A__ = max_input_chars_per_word def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Optional[int]) ->Optional[int]: '''simple docstring''' A__ = list(UpperCAmelCase__) if len(UpperCAmelCase__) > self.max_input_chars_per_word: return [self.unk_token] A__ = 0 A__ = [] while start < len(UpperCAmelCase__): A__ = len(UpperCAmelCase__) A__ = None while start < end: A__ = ''''''.join(chars[start:end]) if substr in self.vocab: A__ = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token) start += 1 else: sub_tokens.append(UpperCAmelCase__) A__ = end return sub_tokens class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = ['''input_ids''', '''attention_mask'''] UpperCAmelCase__ = False def __init__( self : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]="<d>" , UpperCAmelCase__ : Union[str, Any]="</d>" , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Optional[int]="</s>" , UpperCAmelCase__ : Optional[int]="<pad>" , UpperCAmelCase__ : Dict="<unk>" , UpperCAmelCase__ : Optional[Any]="</n>" , UpperCAmelCase__ : List[Any]="</_>" , UpperCAmelCase__ : Optional[Any]="left" , **UpperCAmelCase__ : Union[str, Any] , ) ->Dict: '''simple docstring''' requires_backends(self , ['''jieba''']) super().__init__( bod_token=UpperCAmelCase__ , eod_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , line_token=UpperCAmelCase__ , space_token=UpperCAmelCase__ , padding_side=UpperCAmelCase__ , **UpperCAmelCase__ , ) A__ = bod_token A__ = eod_token A__ = load_vocab(UpperCAmelCase__) A__ = self.encoder[space_token] A__ = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] A__ = collections.OrderedDict(sorted(self.encoder.items() , key=lambda UpperCAmelCase__: x[1])) A__ = {v: k for k, v in self.encoder.items()} A__ = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token) @property def SCREAMING_SNAKE_CASE ( self : List[str]) ->Dict: '''simple docstring''' return self.encoder[self.bod_token] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->List[Any]: '''simple docstring''' return self.encoder[self.eod_token] @property def SCREAMING_SNAKE_CASE ( self : int) ->List[Any]: '''simple docstring''' return self.encoder["\n"] @property def SCREAMING_SNAKE_CASE ( self : int) ->int: '''simple docstring''' return len(self.encoder) def SCREAMING_SNAKE_CASE ( self : Dict) ->str: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder) def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : Any) ->Union[str, Any]: '''simple docstring''' A__ = [] for x in jieba.cut(UpperCAmelCase__ , cut_all=UpperCAmelCase__): output_tokens.extend(self.wordpiece_tokenizer.tokenize(UpperCAmelCase__)) return output_tokens def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Dict) ->List[str]: '''simple docstring''' A__ = [i for i in token_ids if i >= 0] A__ = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(UpperCAmelCase__ , **UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : str , UpperCAmelCase__ : Union[str, Any]) ->Tuple: '''simple docstring''' return token in self.encoder def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : List[str]) ->str: '''simple docstring''' return "".join(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : int) ->Union[str, Any]: '''simple docstring''' return self.encoder.get(UpperCAmelCase__ , self.encoder.get(self.unk_token)) def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : List[str]) ->Any: '''simple docstring''' return self.decoder.get(UpperCAmelCase__ , self.unk_token) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None) ->Tuple[str]: '''simple docstring''' if os.path.isdir(UpperCAmelCase__): A__ = os.path.join( UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) else: A__ = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory A__ = 0 if " " in self.encoder: A__ = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: A__ = self.encoder['''\n'''] del self.encoder["\n"] A__ = collections.OrderedDict(sorted(self.encoder.items() , key=lambda UpperCAmelCase__: x[1])) with open(UpperCAmelCase__ , '''w''' , encoding='''utf-8''') as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ''' Please check that the vocabulary is not corrupted!''') A__ = token_index writer.write(token + '''\n''') index += 1 return (vocab_file,) def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : List[int] = None) ->List[int]: '''simple docstring''' if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def SCREAMING_SNAKE_CASE ( self : Optional[int] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False) ->List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__) if token_ids_a is not None: return [1] + ([0] * len(UpperCAmelCase__)) + [1] + ([0] * len(UpperCAmelCase__)) return [1] + ([0] * len(UpperCAmelCase__))
14
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( lowerCAmelCase ): _a : BigBirdConfig _a : jnp.dtype= jnp.floataa _a : bool= True def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().setup() lowercase : List[str] = nn.Dense(5 ,dtype=self.dtype ) def __call__( self ,*snake_case ,**snake_case ): '''simple docstring''' lowercase : int = super().__call__(*snake_case ,**snake_case ) lowercase : Any = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class __snake_case ( lowerCAmelCase ): _a : List[Any]= FlaxBigBirdForNaturalQuestionsModule def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: def cross_entropy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None ): lowercase : int = logits.shape[-1] lowercase : Dict = (labels[..., None] == jnp.arange(SCREAMING_SNAKE_CASE__ )[None]).astype("""f4""" ) lowercase : Any = jax.nn.log_softmax(SCREAMING_SNAKE_CASE__ , axis=-1 ) lowercase : Optional[Any] = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: lowercase : Any = reduction(SCREAMING_SNAKE_CASE__ ) return loss lowercase : Optional[Any] = partial(SCREAMING_SNAKE_CASE__ , reduction=jnp.mean ) lowercase : Optional[int] = cross_entropy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : Dict = cross_entropy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : int = cross_entropy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : _a : str= "google/bigbird-roberta-base" _a : int= 3000 _a : int= 1_0500 _a : int= 128 _a : int= 3 _a : int= 1 _a : int= 5 # tx_args _a : float= 3E-5 _a : float= 0.0 _a : int= 2_0000 _a : float= 0.00_95 _a : str= "bigbird-roberta-natural-questions" _a : str= "training-expt" _a : str= "data/nq-training.jsonl" _a : str= "data/nq-validation.jsonl" def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' os.makedirs(self.base_dir ,exist_ok=snake_case ) lowercase : Optional[int] = os.path.join(self.base_dir ,self.save_dir ) lowercase : Optional[int] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : _a : int _a : int= 4096 # no dynamic padding on TPUs def __call__( self ,snake_case ): '''simple docstring''' lowercase : int = self.collate_fn(snake_case ) lowercase : Union[str, Any] = jax.tree_util.tree_map(snake_case ,snake_case ) return batch def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase , lowercase : Union[str, Any] = self.fetch_inputs(features["""input_ids"""] ) lowercase : Tuple = { """input_ids""": jnp.array(snake_case ,dtype=jnp.intaa ), """attention_mask""": jnp.array(snake_case ,dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""] ,dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""] ,dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""] ,dtype=jnp.intaa ), } return batch def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Tuple = [self._fetch_inputs(snake_case ) for ids in input_ids] return zip(*snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Union[str, Any] = [1 for _ in range(len(snake_case ) )] while len(snake_case ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None ) -> Any: if seed is not None: lowercase : Optional[int] = dataset.shuffle(seed=SCREAMING_SNAKE_CASE__ ) for i in range(len(SCREAMING_SNAKE_CASE__ ) // batch_size ): lowercase : Optional[Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(SCREAMING_SNAKE_CASE__ ) @partial(jax.pmap , axis_name="""batch""" ) def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> List[Any]: def loss_fn(SCREAMING_SNAKE_CASE__ ): lowercase : List[str] = model_inputs.pop("""start_labels""" ) lowercase : Optional[int] = model_inputs.pop("""end_labels""" ) lowercase : str = model_inputs.pop("""pooled_labels""" ) lowercase : Union[str, Any] = state.apply_fn(**SCREAMING_SNAKE_CASE__ , params=SCREAMING_SNAKE_CASE__ , dropout_rng=SCREAMING_SNAKE_CASE__ , train=SCREAMING_SNAKE_CASE__ ) lowercase , lowercase , lowercase : List[str] = outputs return state.loss_fn( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) lowercase , lowercase : int = jax.random.split(SCREAMING_SNAKE_CASE__ ) lowercase : Dict = jax.value_and_grad(SCREAMING_SNAKE_CASE__ ) lowercase , lowercase : Union[str, Any] = grad_fn(state.params ) lowercase : List[Any] = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) lowercase : List[Any] = jax.lax.pmean(SCREAMING_SNAKE_CASE__ , """batch""" ) lowercase : str = state.apply_gradients(grads=SCREAMING_SNAKE_CASE__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="""batch""" ) def _snake_case( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> List[str]: lowercase : int = model_inputs.pop("""start_labels""" ) lowercase : Dict = model_inputs.pop("""end_labels""" ) lowercase : Optional[Any] = model_inputs.pop("""pooled_labels""" ) lowercase : Optional[int] = state.apply_fn(**SCREAMING_SNAKE_CASE__ , params=state.params , train=SCREAMING_SNAKE_CASE__ ) lowercase , lowercase , lowercase : List[Any] = outputs lowercase : Dict = state.loss_fn(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : str = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) return metrics class __snake_case ( train_state.TrainState ): _a : Callable= struct.field(pytree_node=lowerCAmelCase ) @dataclass class __snake_case : _a : Args _a : Callable _a : Callable _a : Callable _a : Callable _a : wandb _a : Callable= None def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case ,snake_case=None ): '''simple docstring''' lowercase : Tuple = model.params lowercase : Any = TrainState.create( apply_fn=model.__call__ ,params=snake_case ,tx=snake_case ,loss_fn=snake_case ,) if ckpt_dir is not None: lowercase , lowercase , lowercase , lowercase , lowercase : Tuple = restore_checkpoint(snake_case ,snake_case ) lowercase : List[str] = { """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } lowercase , lowercase : Tuple = build_tx(**snake_case ) lowercase : str = train_state.TrainState( step=snake_case ,apply_fn=model.__call__ ,params=snake_case ,tx=snake_case ,opt_state=snake_case ,) lowercase : Any = args lowercase : Optional[Any] = data_collator lowercase : List[str] = lr lowercase : str = params lowercase : Tuple = jax_utils.replicate(snake_case ) return state def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case ): '''simple docstring''' lowercase : Dict = self.args lowercase : Optional[Any] = len(snake_case ) // args.batch_size lowercase : int = jax.random.PRNGKey(0 ) lowercase : List[str] = jax.random.split(snake_case ,jax.device_count() ) for epoch in range(args.max_epochs ): lowercase : List[Any] = jnp.array(0 ,dtype=jnp.floataa ) lowercase : List[str] = get_batched_dataset(snake_case ,args.batch_size ,seed=snake_case ) lowercase : int = 0 for batch in tqdm(snake_case ,total=snake_case ,desc=f"Running EPOCH-{epoch}" ): lowercase : Dict = self.data_collator(snake_case ) lowercase , lowercase , lowercase : Optional[int] = self.train_step_fn(snake_case ,snake_case ,**snake_case ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: lowercase : Optional[Any] = jax_utils.unreplicate(state.step ) lowercase : List[str] = running_loss.item() / i lowercase : List[str] = self.scheduler_fn(state_step - 1 ) lowercase : int = self.evaluate(snake_case ,snake_case ) lowercase : Tuple = { """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(snake_case ) ) self.logger.log(snake_case ,commit=snake_case ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f"-e{epoch}-s{i}" ,state=snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : List[str] = get_batched_dataset(snake_case ,self.args.batch_size ) lowercase : Any = len(snake_case ) // self.args.batch_size lowercase : List[Any] = jnp.array(0 ,dtype=jnp.floataa ) lowercase : Optional[int] = 0 for batch in tqdm(snake_case ,total=snake_case ,desc="""Evaluating ... """ ): lowercase : Tuple = self.data_collator(snake_case ) lowercase : Optional[int] = self.val_step_fn(snake_case ,**snake_case ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : str = jax_utils.unreplicate(snake_case ) print(f"SAVING CHECKPOINT IN {save_dir}" ,end=""" ... """ ) self.model_save_fn(snake_case ,params=state.params ) with open(os.path.join(snake_case ,"""opt_state.msgpack""" ) ,"""wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args ,os.path.join(snake_case ,"""args.joblib""" ) ) joblib.dump(self.data_collator ,os.path.join(snake_case ,"""data_collator.joblib""" ) ) with open(os.path.join(snake_case ,"""training_state.json""" ) ,"""w""" ) as f: json.dump({"""step""": state.step.item()} ,snake_case ) print("""DONE""" ) def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: print(f"RESTORING CHECKPOINT FROM {save_dir}" , end=""" ... """ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , """flax_model.msgpack""" ) , """rb""" ) as f: lowercase : str = from_bytes(state.params , f.read() ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , """opt_state.msgpack""" ) , """rb""" ) as f: lowercase : Optional[int] = from_bytes(state.opt_state , f.read() ) lowercase : Optional[Any] = joblib.load(os.path.join(SCREAMING_SNAKE_CASE__ , """args.joblib""" ) ) lowercase : int = joblib.load(os.path.join(SCREAMING_SNAKE_CASE__ , """data_collator.joblib""" ) ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , """training_state.json""" ) , """r""" ) as f: lowercase : Tuple = json.load(SCREAMING_SNAKE_CASE__ ) lowercase : Tuple = training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[Any]: lowercase : List[str] = num_train_steps - warmup_steps lowercase : Dict = optax.linear_schedule(init_value=SCREAMING_SNAKE_CASE__ , end_value=SCREAMING_SNAKE_CASE__ , transition_steps=SCREAMING_SNAKE_CASE__ ) lowercase : List[str] = optax.linear_schedule(init_value=SCREAMING_SNAKE_CASE__ , end_value=1e-7 , transition_steps=SCREAMING_SNAKE_CASE__ ) lowercase : Tuple = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: def weight_decay_mask(SCREAMING_SNAKE_CASE__ ): lowercase : List[Any] = traverse_util.flatten_dict(SCREAMING_SNAKE_CASE__ ) lowercase : List[Any] = {k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(SCREAMING_SNAKE_CASE__ ) lowercase : List[str] = scheduler_fn(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : List[str] = optax.adamw(learning_rate=SCREAMING_SNAKE_CASE__ , weight_decay=SCREAMING_SNAKE_CASE__ , mask=SCREAMING_SNAKE_CASE__ ) return tx, lr
20
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : Union[str, Any] = { "configuration_roformer": ["ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoFormerConfig", "RoFormerOnnxConfig"], "tokenization_roformer": ["RoFormerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = ["RoFormerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = [ "ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "RoFormerForCausalLM", "RoFormerForMaskedLM", "RoFormerForMultipleChoice", "RoFormerForQuestionAnswering", "RoFormerForSequenceClassification", "RoFormerForTokenClassification", "RoFormerLayer", "RoFormerModel", "RoFormerPreTrainedModel", "load_tf_weights_in_roformer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Optional[int] = [ "TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRoFormerForCausalLM", "TFRoFormerForMaskedLM", "TFRoFormerForMultipleChoice", "TFRoFormerForQuestionAnswering", "TFRoFormerForSequenceClassification", "TFRoFormerForTokenClassification", "TFRoFormerLayer", "TFRoFormerModel", "TFRoFormerPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : List[str] = [ "FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxRoFormerForMaskedLM", "FlaxRoFormerForMultipleChoice", "FlaxRoFormerForQuestionAnswering", "FlaxRoFormerForSequenceClassification", "FlaxRoFormerForTokenClassification", "FlaxRoFormerModel", "FlaxRoFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys _lowerCAmelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
202
"""simple docstring""" import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu _lowerCAmelCase : Union[str, Any] = get_tests_dir() + "/test_data/fsmt/fsmt_val_data.json" with io.open(filename, "r", encoding="utf-8") as f: _lowerCAmelCase : Tuple = json.load(f) @require_torch class UpperCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self : str , A : Union[str, Any] ): return FSMTTokenizer.from_pretrained(A ) def snake_case_ ( self : Union[str, Any] , A : Union[str, Any] ): _UpperCAmelCase : List[Any] = FSMTForConditionalGeneration.from_pretrained(A ).to(A ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ["en-ru", 26.0], ["ru-en", 22.0], ["en-de", 22.0], ["de-en", 29.0], ] ) @slow def snake_case_ ( self : Any , A : Dict , A : List[str] ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality _UpperCAmelCase : Any = f'facebook/wmt19-{pair}' _UpperCAmelCase : Dict = self.get_tokenizer(A ) _UpperCAmelCase : Optional[int] = self.get_model(A ) _UpperCAmelCase : int = bleu_data[pair]["src"] _UpperCAmelCase : Optional[int] = bleu_data[pair]["tgt"] _UpperCAmelCase : List[str] = tokenizer(A , return_tensors="pt" , truncation=A , padding="longest" ).to(A ) _UpperCAmelCase : List[str] = model.generate( input_ids=batch.input_ids , num_beams=8 , ) _UpperCAmelCase : Any = tokenizer.batch_decode( A , skip_special_tokens=A , clean_up_tokenization_spaces=A ) _UpperCAmelCase : Any = calculate_bleu(A , A ) print(A ) self.assertGreaterEqual(scores["bleu"] , A )
202
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device __A = False class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> int: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = 'A painting of a squirrel eating a burger ' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = 'A painting of a squirrel eating a burger ' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
90
'''simple docstring''' def a ( ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ :Optional[int] = [] UpperCamelCase__ :int = 1 while len(__a ) < 1e6: constant.append(str(__a ) ) i += 1 UpperCamelCase__ :Union[str, Any] = ''''''.join(__a ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[99999] ) * int(constant[999999] ) ) if __name__ == "__main__": print(solution())
97
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase : int = logging.get_logger(__name__) _lowercase : Optional[Any] = { "tiiuae/falcon-40b": "https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json", "tiiuae/falcon-7b": "https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json", } class UpperCamelCase__( snake_case__ ): __magic_name__ : Optional[Any] = "falcon" __magic_name__ : List[Any] = ["past_key_values"] def __init__( self : int , lowerCAmelCase : Any=65024 , lowerCAmelCase : int=4544 , lowerCAmelCase : List[Any]=32 , lowerCAmelCase : Any=71 , lowerCAmelCase : str=1E-5 , lowerCAmelCase : Any=0.02 , lowerCAmelCase : Optional[Any]=True , lowerCAmelCase : str=0.0 , lowerCAmelCase : List[Any]=0.0 , lowerCAmelCase : int=None , lowerCAmelCase : Optional[Any]=False , lowerCAmelCase : List[Any]=False , lowerCAmelCase : Dict=True , lowerCAmelCase : Optional[Any]=True , lowerCAmelCase : Tuple=False , lowerCAmelCase : Optional[Any]=11 , lowerCAmelCase : Dict=11 , **lowerCAmelCase : Union[str, Any] , )-> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size # Backward compatibility with n_embed kwarg UpperCAmelCase = kwargs.pop('''n_embed''' , UpperCAmelCase_ ) UpperCAmelCase = hidden_size if n_embed is None else n_embed UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = bos_token_id UpperCAmelCase = eos_token_id UpperCAmelCase = num_attention_heads if num_kv_heads is None else num_kv_heads UpperCAmelCase = alibi UpperCAmelCase = new_decoder_architecture UpperCAmelCase = multi_query # Ignored when new_decoder_architecture is True UpperCAmelCase = parallel_attn UpperCAmelCase = bias super().__init__(bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) @property def a__( self : Any )-> Dict: """simple docstring""" return self.hidden_size // self.num_attention_heads @property def a__( self : str )-> Union[str, Any]: """simple docstring""" return not self.alibi
365
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase : Union[str, Any] = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """PoolFormerOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : List[Any] = ["""PoolFormerFeatureExtractor"""] _lowercase : Any = ["""PoolFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Tuple = [ """POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PoolFormerForImageClassification""", """PoolFormerModel""", """PoolFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _lowercase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
91
0
'''simple docstring''' import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel _snake_case = logging.getLogger(__name__) def _A ( snake_case , snake_case ) -> List[Any]: # save results if os.path.exists(snake_case ): if os.path.exists(os.path.join(snake_case , "config.json" ) ) and os.path.isfile( os.path.join(snake_case , "config.json" ) ): os.remove(os.path.join(snake_case , "config.json" ) ) if os.path.exists(os.path.join(snake_case , "pytorch_model.bin" ) ) and os.path.isfile( os.path.join(snake_case , "pytorch_model.bin" ) ): os.remove(os.path.join(snake_case , "pytorch_model.bin" ) ) else: os.makedirs(snake_case ) model.save_pretrained(snake_case ) def _A ( snake_case , snake_case=False ) -> int: _lowercase : Union[str, Any] = 2 if unlogit: _lowercase : Optional[Any] = torch.pow(snake_case , snake_case ) _lowercase : List[Any] = p * torch.log(snake_case ) _lowercase : str = 0 return -plogp.sum(dim=-1 ) def _A ( snake_case ) -> List[Any]: logger.info("lv, h >\t" + "\t".join(F'''{x + 1}''' for x in range(len(snake_case ) ) ) ) for row in range(len(snake_case ) ): if tensor.dtype != torch.long: logger.info(F'''layer {row + 1}:\t''' + "\t".join(F'''{x:.5f}''' for x in tensor[row].cpu().data ) ) else: logger.info(F'''layer {row + 1}:\t''' + "\t".join(F'''{x:d}''' for x in tensor[row].cpu().data ) ) def _A ( snake_case , snake_case , snake_case , snake_case=True , snake_case=True , snake_case=None , snake_case=False ) -> Optional[int]: _lowercase , _lowercase : Union[str, Any] = model.config.num_hidden_layers, model.config.num_attention_heads _lowercase : Optional[int] = torch.zeros(snake_case , snake_case ).to(args.device ) _lowercase : str = torch.zeros(snake_case , snake_case ).to(args.device ) if head_mask is None: _lowercase : Any = torch.ones(snake_case , snake_case ).to(args.device ) head_mask.requires_grad_(requires_grad=snake_case ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: _lowercase : int = None _lowercase : List[str] = 0.0 _lowercase : str = 0.0 for step, inputs in enumerate(tqdm(snake_case , desc="Iteration" , disable=args.local_rank not in [-1, 0] ) ): _lowercase : Dict = tuple(t.to(args.device ) for t in inputs ) ((_lowercase) , ) : Any = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) _lowercase : str = model(snake_case , labels=snake_case , head_mask=snake_case ) # (loss), lm_logits, presents, (all hidden_states), (attentions) _lowercase , _lowercase , _lowercase : Optional[int] = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(snake_case ): _lowercase : Optional[int] = entropy(attn.detach() , snake_case ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(snake_case ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: _lowercase : List[str] = 2 _lowercase : Dict = torch.pow(torch.pow(snake_case , snake_case ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1E-20 if not args.dont_normalize_global_importance: _lowercase : str = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("Attention entropies" ) print_ad_tensor(snake_case ) if compute_importance: logger.info("Head importance scores" ) print_ad_tensor(snake_case ) logger.info("Head ranked by importance scores" ) _lowercase : Any = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) _lowercase : Union[str, Any] = torch.arange( head_importance.numel() , device=args.device ) _lowercase : Optional[Any] = head_ranks.view_as(snake_case ) print_ad_tensor(snake_case ) return attn_entropy, head_importance, total_loss def _A ( snake_case , snake_case , snake_case ) -> Optional[Any]: _lowercase , _lowercase , _lowercase : Union[str, Any] = compute_heads_importance(snake_case , snake_case , snake_case , compute_entropy=snake_case ) _lowercase : int = 1 / loss # instead of downsteam score use the LM loss logger.info("Pruning: original score: %f, threshold: %f" , snake_case , original_score * args.masking_threshold ) _lowercase : List[Any] = torch.ones_like(snake_case ) _lowercase : Dict = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) _lowercase : Union[str, Any] = original_score while current_score >= original_score * args.masking_threshold: _lowercase : Any = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads _lowercase : Dict = float("Inf" ) _lowercase : Union[str, Any] = head_importance.view(-1 ).sort()[1] if len(snake_case ) <= num_to_mask: print("BREAK BY num_to_mask" ) break # mask heads _lowercase : List[str] = current_heads_to_mask[:num_to_mask] logger.info("Heads to mask: %s" , str(current_heads_to_mask.tolist() ) ) _lowercase : int = new_head_mask.view(-1 ) _lowercase : Union[str, Any] = 0.0 _lowercase : Dict = new_head_mask.view_as(snake_case ) _lowercase : str = new_head_mask.clone().detach() print_ad_tensor(snake_case ) # Compute metric and head importance again _lowercase , _lowercase , _lowercase : Any = compute_heads_importance( snake_case , snake_case , snake_case , compute_entropy=snake_case , head_mask=snake_case ) _lowercase : str = 1 / loss logger.info( "Masking: current score: %f, remaining heads %d (%.1f percents)" , snake_case , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 1_00 , ) logger.info("Final head mask" ) print_ad_tensor(snake_case ) np.save(os.path.join(args.output_dir , "head_mask.npy" ) , head_mask.detach().cpu().numpy() ) return head_mask def _A ( snake_case , snake_case , snake_case , snake_case ) -> Any: _lowercase : List[Any] = datetime.now() _lowercase , _lowercase , _lowercase : List[Any] = compute_heads_importance( snake_case , snake_case , snake_case , compute_entropy=snake_case , compute_importance=snake_case , head_mask=snake_case ) _lowercase : Tuple = 1 / loss _lowercase : List[Any] = datetime.now() - before_time _lowercase : int = sum(p.numel() for p in model.parameters() ) _lowercase : str = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(snake_case ) ) } for k, v in heads_to_prune.items(): if isinstance(snake_case , snake_case ): _lowercase : Optional[Any] = [ v, ] assert sum(len(snake_case ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(snake_case ) _lowercase : List[str] = sum(p.numel() for p in model.parameters() ) _lowercase : int = datetime.now() _lowercase , _lowercase , _lowercase : Any = compute_heads_importance( snake_case , snake_case , snake_case , compute_entropy=snake_case , compute_importance=snake_case , head_mask=snake_case , actually_pruned=snake_case , ) _lowercase : List[Any] = 1 / loss _lowercase : int = datetime.now() - before_time logger.info( "Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)" , snake_case , snake_case , pruned_num_params / original_num_params * 1_00 , ) logger.info("Pruning: score with masking: %f score with pruning: %f" , snake_case , snake_case ) logger.info("Pruning: speed ratio (original timing / new timing): %f percents" , original_time / new_time * 1_00 ) save_model(snake_case , args.output_dir ) def _A ( ) -> int: _lowercase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--data_dir" , default=snake_case , type=snake_case , required=snake_case , help="The input data dir. Should contain the .tsv files (or other data files) for the task." , ) parser.add_argument( "--model_name_or_path" , default=snake_case , type=snake_case , required=snake_case , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--output_dir" , default=snake_case , type=snake_case , required=snake_case , help="The output directory where the model predictions and checkpoints will be written." , ) # Other parameters parser.add_argument( "--config_name" , default="" , type=snake_case , help="Pretrained config name or path if not the same as model_name_or_path" , ) parser.add_argument( "--tokenizer_name" , default="" , type=snake_case , help="Pretrained tokenizer name or path if not the same as model_name_or_path" , ) parser.add_argument( "--cache_dir" , default=snake_case , type=snake_case , help="Where do you want to store the pre-trained models downloaded from s3" , ) parser.add_argument( "--data_subset" , type=snake_case , default=-1 , help="If > 0: limit the data to a subset of data_subset instances." ) parser.add_argument( "--overwrite_output_dir" , action="store_true" , help="Whether to overwrite data in output directory" ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--dont_normalize_importance_by_layer" , action="store_true" , help="Don't normalize importance score by layers" ) parser.add_argument( "--dont_normalize_global_importance" , action="store_true" , help="Don't normalize all importance scores between 0 and 1" , ) parser.add_argument( "--try_masking" , action="store_true" , help="Whether to try to mask head until a threshold of accuracy." ) parser.add_argument( "--masking_threshold" , default=0.9 , type=snake_case , help="masking threshold in term of metrics (stop masking when metric < threshold * original metric value)." , ) parser.add_argument( "--masking_amount" , default=0.1 , type=snake_case , help="Amount to heads to masking at each masking step." ) parser.add_argument("--metric_name" , default="acc" , type=snake_case , help="Metric to use for head masking." ) parser.add_argument( "--max_seq_length" , default=1_28 , type=snake_case , help=( "The maximum total input sequence length after WordPiece tokenization. \n" "Sequences longer than this will be truncated, sequences shorter padded." ) , ) parser.add_argument("--batch_size" , default=1 , type=snake_case , help="Batch size." ) parser.add_argument("--seed" , type=snake_case , default=42 ) parser.add_argument("--local_rank" , type=snake_case , default=-1 , help="local_rank for distributed training on gpus" ) parser.add_argument("--no_cuda" , action="store_true" , help="Whether not to use CUDA when available" ) parser.add_argument("--server_ip" , type=snake_case , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=snake_case , default="" , help="Can be used for distant debugging." ) _lowercase : Optional[int] = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=snake_case ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: _lowercase : Any = torch.device("cuda" if torch.cuda.is_available() and not args.no_cuda else "cpu" ) _lowercase : Optional[int] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) _lowercase : List[Any] = torch.device("cuda" , args.local_rank ) _lowercase : Dict = 1 torch.distributed.init_process_group(backend="nccl" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) _lowercase : List[Any] = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: _lowercase : str = nn.parallel.DistributedDataParallel( snake_case , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=snake_case ) elif args.n_gpu > 1: _lowercase : Dict = nn.DataParallel(snake_case ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=snake_case ) torch.save(snake_case , os.path.join(args.output_dir , "run_args.bin" ) ) logger.info("Training/evaluation parameters %s" , snake_case ) # Prepare dataset _lowercase : Optional[Any] = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) _lowercase : List[str] = (torch.from_numpy(snake_case ),) _lowercase : Dict = TensorDataset(*snake_case ) _lowercase : List[Any] = RandomSampler(snake_case ) _lowercase : str = DataLoader(snake_case , sampler=snake_case , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(snake_case , snake_case , snake_case ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: _lowercase : int = mask_heads(snake_case , snake_case , snake_case ) prune_heads(snake_case , snake_case , snake_case , snake_case ) if __name__ == "__main__": main()
250
'''simple docstring''' import argparse import os import re _snake_case = 'src/transformers' # Pattern that looks at the indentation in a line. _snake_case = re.compile(r'^(\s*)\S') # Pattern that matches `"key":" and puts `key` in group 0. _snake_case = re.compile(r'^\s*"([^"]+)":') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _snake_case = re.compile(r'^\s*_import_structure\["([^"]+)"\]') # Pattern that matches `"key",` and puts `key` in group 0. _snake_case = re.compile(r'^\s*"([^"]+)",\s*$') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _snake_case = re.compile(r'\[([^\]]+)\]') def _A ( snake_case ) -> str: _lowercase : Union[str, Any] = _re_indent.search(snake_case ) return "" if search is None else search.groups()[0] def _A ( snake_case , snake_case="" , snake_case=None , snake_case=None ) -> Optional[int]: _lowercase : List[str] = 0 _lowercase : str = code.split("\n" ) if start_prompt is not None: while not lines[index].startswith(snake_case ): index += 1 _lowercase : Optional[int] = ["\n".join(lines[:index] )] else: _lowercase : Dict = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). _lowercase : Any = [lines[index]] index += 1 while index < len(snake_case ) and (end_prompt is None or not lines[index].startswith(snake_case )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(snake_case ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + " " ): current_block.append(lines[index] ) blocks.append("\n".join(snake_case ) ) if index < len(snake_case ) - 1: _lowercase : int = [lines[index + 1]] index += 1 else: _lowercase : Optional[int] = [] else: blocks.append("\n".join(snake_case ) ) _lowercase : Optional[Any] = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(snake_case ) > 0: blocks.append("\n".join(snake_case ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(snake_case ): blocks.append("\n".join(lines[index:] ) ) return blocks def _A ( snake_case ) -> Optional[int]: def _inner(snake_case ): return key(snake_case ).lower().replace("_" , "" ) return _inner def _A ( snake_case , snake_case=None ) -> List[str]: # If no key is provided, we use a noop. def noop(snake_case ): return x if key is None: _lowercase : Optional[int] = noop # Constants are all uppercase, they go first. _lowercase : Dict = [obj for obj in objects if key(snake_case ).isupper()] # Classes are not all uppercase but start with a capital, they go second. _lowercase : int = [obj for obj in objects if key(snake_case )[0].isupper() and not key(snake_case ).isupper()] # Functions begin with a lowercase, they go last. _lowercase : Dict = [obj for obj in objects if not key(snake_case )[0].isupper()] _lowercase : Union[str, Any] = ignore_underscore(snake_case ) return sorted(snake_case , key=snake_case ) + sorted(snake_case , key=snake_case ) + sorted(snake_case , key=snake_case ) def _A ( snake_case ) -> List[Any]: # This inner function sort imports between [ ]. def _replace(snake_case ): _lowercase : Optional[Any] = match.groups()[0] if "," not in imports: return F'''[{imports}]''' _lowercase : str = [part.strip().replace("\"" , "" ) for part in imports.split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: _lowercase : Optional[int] = keys[:-1] return "[" + ", ".join([F'''"{k}"''' for k in sort_objects(snake_case )] ) + "]" _lowercase : Tuple = import_statement.split("\n" ) if len(snake_case ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. _lowercase : Union[str, Any] = 2 if lines[1].strip() == "[" else 1 _lowercase : Optional[int] = [(i, _re_strip_line.search(snake_case ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] _lowercase : Any = sort_objects(snake_case , key=lambda snake_case : x[1] ) _lowercase : Tuple = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(snake_case ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: _lowercase : Dict = _re_bracket_content.sub(_replace , lines[1] ) else: _lowercase : Optional[Any] = [part.strip().replace("\"" , "" ) for part in lines[1].split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: _lowercase : Optional[int] = keys[:-1] _lowercase : Optional[Any] = get_indent(lines[1] ) + ", ".join([F'''"{k}"''' for k in sort_objects(snake_case )] ) return "\n".join(snake_case ) else: # Finally we have to deal with imports fitting on one line _lowercase : Optional[Any] = _re_bracket_content.sub(_replace , snake_case ) return import_statement def _A ( snake_case , snake_case=True ) -> Dict: with open(snake_case , encoding="utf-8" ) as f: _lowercase : Dict = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 _lowercase : Optional[Any] = split_code_in_indented_blocks( snake_case , start_prompt="_import_structure = {" , end_prompt="if TYPE_CHECKING:" ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(snake_case ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. _lowercase : Dict = main_blocks[block_idx] _lowercase : Union[str, Any] = block.split("\n" ) # Get to the start of the imports. _lowercase : int = 0 while line_idx < len(snake_case ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: _lowercase : Optional[Any] = len(snake_case ) else: line_idx += 1 if line_idx >= len(snake_case ): continue # Ignore beginning and last line: they don't contain anything. _lowercase : Any = "\n".join(block_lines[line_idx:-1] ) _lowercase : int = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. _lowercase : Optional[int] = split_code_in_indented_blocks(snake_case , indent_level=snake_case ) # We have two categories of import key: list or _import_structure[key].append/extend _lowercase : Union[str, Any] = _re_direct_key if "_import_structure = {" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. _lowercase : str = [(pattern.search(snake_case ).groups()[0] if pattern.search(snake_case ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. _lowercase : List[str] = [(i, key) for i, key in enumerate(snake_case ) if key is not None] _lowercase : Tuple = [x[0] for x in sorted(snake_case , key=lambda snake_case : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. _lowercase : Any = 0 _lowercase : str = [] for i in range(len(snake_case ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: _lowercase : Tuple = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(snake_case ) count += 1 # And we put our main block back together with its first and last line. _lowercase : int = "\n".join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(snake_case ): if check_only: return True else: print(F'''Overwriting {file}.''' ) with open(snake_case , "w" , encoding="utf-8" ) as f: f.write("\n".join(snake_case ) ) def _A ( snake_case=True ) -> str: _lowercase : List[Any] = [] for root, _, files in os.walk(snake_case ): if "__init__.py" in files: _lowercase : Tuple = sort_imports(os.path.join(snake_case , "__init__.py" ) , check_only=snake_case ) if result: _lowercase : Any = [os.path.join(snake_case , "__init__.py" )] if len(snake_case ) > 0: raise ValueError(F'''Would overwrite {len(snake_case )} files, run `make style`.''' ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') _snake_case = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
250
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": __UpperCamelCase : Optional[int] = pd.read_csv("sample_data.csv", header=None) __UpperCamelCase : Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column __UpperCamelCase : Tuple = df.iloc[:, 1:2] __UpperCamelCase : Tuple = actual_data.values.reshape(len_data, 1) __UpperCamelCase : str = MinMaxScaler().fit_transform(actual_data) __UpperCamelCase : List[str] = 10 __UpperCamelCase : Any = 5 __UpperCamelCase : Optional[Any] = 20 __UpperCamelCase : List[str] = len_data - periods * look_back __UpperCamelCase : str = actual_data[:division] __UpperCamelCase : int = actual_data[division - look_back :] __UpperCamelCase : List[str] = [], [] __UpperCamelCase : Union[str, Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) __UpperCamelCase : List[Any] = np.array(train_x) __UpperCamelCase : Tuple = np.array(test_x) __UpperCamelCase : Any = np.array([list(i.ravel()) for i in train_y]) __UpperCamelCase : List[Any] = np.array([list(i.ravel()) for i in test_y]) __UpperCamelCase : Union[str, Any] = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") __UpperCamelCase : Tuple = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) __UpperCamelCase : Optional[int] = model.predict(x_test)
355
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __UpperCamelCase : List[Any] = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" inspect_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = path + '''.py''' assert script_name in os.listdir(SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def _a ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" inspect_metric(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : int = path + '''.py''' assert script_name in os.listdir(SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : str ): """simple docstring""" UpperCamelCase__ : int = get_dataset_config_info(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _a ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : str ): """simple docstring""" with pytest.raises(SCREAMING_SNAKE_CASE ): get_dataset_config_info(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def _a ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" UpperCamelCase__ : List[Any] = get_dataset_config_names(SCREAMING_SNAKE_CASE ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" UpperCamelCase__ : Optional[int] = get_dataset_infos(SCREAMING_SNAKE_CASE ) assert list(infos.keys() ) == expected_configs UpperCamelCase__ : List[str] = expected_configs[0] assert expected_config in infos UpperCamelCase__ : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" UpperCamelCase__ : Optional[Any] = get_dataset_infos(SCREAMING_SNAKE_CASE ) assert expected_config in infos UpperCamelCase__ : Optional[int] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _a ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" with pytest.raises(SCREAMING_SNAKE_CASE ): get_dataset_split_names(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE )
51
0
import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def a_ ( _A , _A ) -> str: """simple docstring""" snake_case__ = args.log_outputs snake_case__ = '_'.join(args.dataset.split('/' ) + [args.config, args.split] ) # load metric snake_case__ = load_metric('wer' ) snake_case__ = load_metric('cer' ) # compute metrics snake_case__ = wer.compute(references=result['target'] , predictions=result['prediction'] ) snake_case__ = cer.compute(references=result['target'] , predictions=result['prediction'] ) # print & log results snake_case__ = f'''WER: {wer_result}\nCER: {cer_result}''' print(_A ) with open(f'''{dataset_id}_eval_results.txt''' , 'w' ) as f: f.write(_A ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: snake_case__ = f'''log_{dataset_id}_predictions.txt''' snake_case__ = f'''log_{dataset_id}_targets.txt''' with open(_A , 'w' ) as p, open(_A , 'w' ) as t: # mapping function to write output def write_to_file(_A , _A ): p.write(f'''{i}''' + '\n' ) p.write(batch['prediction'] + '\n' ) t.write(f'''{i}''' + '\n' ) t.write(batch['target'] + '\n' ) result.map(_A , with_indices=_A ) def a_ ( _A ) -> str: """simple docstring""" snake_case__ = '[,?.!\-\;\:"“%‘”�—’…–]' # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training snake_case__ = re.sub(_A , '' , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! snake_case__ = ['\n\n', '\n', ' ', ' '] for t in token_sequences_to_ignore: snake_case__ = ' '.join(text.split(_A ) ) return text def a_ ( _A ) -> Optional[Any]: """simple docstring""" # load dataset snake_case__ = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=_A ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor snake_case__ = AutoFeatureExtractor.from_pretrained(args.model_id ) snake_case__ = feature_extractor.sampling_rate # resample audio snake_case__ = dataset.cast_column('audio' , Audio(sampling_rate=_A ) ) # load eval pipeline if args.device is None: snake_case__ = 0 if torch.cuda.is_available() else -1 snake_case__ = pipeline('automatic-speech-recognition' , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(_A ): snake_case__ = asr( batch['audio']['array'] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) snake_case__ = prediction['text'] snake_case__ = normalize_text(batch['sentence'] ) return batch # run inference on all examples snake_case__ = dataset.map(_A , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(_A , _A ) if __name__ == "__main__": __UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--model_id""", type=str, required=True, help="""Model identifier. Should be loadable with 🤗 Transformers""" ) parser.add_argument( """--dataset""", type=str, required=True, help="""Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets""", ) parser.add_argument( """--config""", type=str, required=True, help="""Config of the dataset. *E.g.* `'en'` for Common Voice""" ) parser.add_argument("""--split""", type=str, required=True, help="""Split of the dataset. *E.g.* `'test'`""") parser.add_argument( """--chunk_length_s""", type=float, default=None, help="""Chunk length in seconds. Defaults to 5 seconds.""" ) parser.add_argument( """--stride_length_s""", type=float, default=None, help="""Stride of the audio chunks. Defaults to 1 second.""" ) parser.add_argument( """--log_outputs""", action="""store_true""", help="""If defined, write outputs to log file for analysis.""" ) parser.add_argument( """--device""", type=int, default=None, help="""The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.""", ) __UpperCamelCase : Optional[int] = parser.parse_args() main(args)
307
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : Any = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys __UpperCamelCase : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
307
1
def lowerCAmelCase_ ( _lowercase : int = 1000) -> int: """simple docstring""" return sum(e for e in range(3 , _lowercase) if e % 3 == 0 or e % 5 == 0) if __name__ == "__main__": print(f'{solution() = }')
361
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _lowercase : Union[str, Any] =16 _lowercase : Dict =32 def lowerCAmelCase_ ( _lowercase : Accelerator , _lowercase : int = 16 , _lowercase : str = "bert-base-cased") -> Union[str, Any]: """simple docstring""" a__ : Union[str, Any] = AutoTokenizer.from_pretrained(_lowercase) a__ : Any = load_dataset("""glue""" , """mrpc""") def tokenize_function(_lowercase : str): # max_length=None => use the model max length (it's actually the default) a__ : List[str] = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=_lowercase , max_length=_lowercase) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset a__ : Any = datasets.map( _lowercase , batched=_lowercase , remove_columns=["""idx""", """sentence1""", """sentence2"""] , load_from_cache_file=_lowercase) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library a__ : Tuple = tokenized_datasets.rename_column("""label""" , """labels""") def collate_fn(_lowercase : Optional[Any]): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(_lowercase , padding="""max_length""" , max_length=128 , return_tensors="""pt""") return tokenizer.pad(_lowercase , padding="""longest""" , return_tensors="""pt""") # Instantiate dataloaders. a__ : int = DataLoader( tokenized_datasets["""train"""] , shuffle=_lowercase , collate_fn=_lowercase , batch_size=_lowercase) a__ : Union[str, Any] = DataLoader( tokenized_datasets["""validation"""] , shuffle=_lowercase , collate_fn=_lowercase , batch_size=_lowercase) return train_dataloader, eval_dataloader def lowerCAmelCase_ ( _lowercase : Any , _lowercase : Dict) -> int: """simple docstring""" # Initialize accelerator a__ : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs a__ : Union[str, Any] = config["""lr"""] a__ : List[str] = int(config["""num_epochs"""]) a__ : List[str] = int(config["""seed"""]) a__ : Tuple = int(config["""batch_size"""]) a__ : int = args.model_name_or_path set_seed(_lowercase) a__ , a__ : int = get_dataloaders(_lowercase , _lowercase , _lowercase) # Instantiate the model (we build the model here so that the seed also control new weights initialization) a__ : Optional[Any] = AutoModelForSequenceClassification.from_pretrained(_lowercase , return_dict=_lowercase) # Instantiate optimizer a__ : int = ( AdamW if accelerator.state.deepspeed_plugin is None or """optimizer""" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) a__ : Optional[Any] = optimizer_cls(params=model.parameters() , lr=_lowercase) if accelerator.state.deepspeed_plugin is not None: a__ : Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ """gradient_accumulation_steps""" ] else: a__ : List[str] = 1 a__ : List[Any] = (len(_lowercase) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): a__ : Dict = get_linear_schedule_with_warmup( optimizer=_lowercase , num_warmup_steps=0 , num_training_steps=_lowercase , ) else: a__ : Dict = DummyScheduler(_lowercase , total_num_steps=_lowercase , warmup_num_steps=0) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. a__ , a__ , a__ , a__ , a__ : List[Any] = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase) # We need to keep track of how many total steps we have iterated over a__ : Optional[int] = 0 # We also need to keep track of the stating epoch so files are named properly a__ : Optional[int] = 0 # Now we train the model a__ : Tuple = evaluate.load("""glue""" , """mrpc""") a__ : List[Any] = 0 a__ : Tuple = {} for epoch in range(_lowercase , _lowercase): model.train() for step, batch in enumerate(_lowercase): a__ : Union[str, Any] = model(**_lowercase) a__ : Tuple = outputs.loss a__ : Any = loss / gradient_accumulation_steps accelerator.backward(_lowercase) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() a__ : int = 0 for step, batch in enumerate(_lowercase): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device) with torch.no_grad(): a__ : str = model(**_lowercase) a__ : Union[str, Any] = outputs.logits.argmax(dim=-1) # It is slightly faster to call this once, than multiple times a__ , a__ : Optional[int] = accelerator.gather( (predictions, batch["""labels"""])) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(_lowercase) - 1: a__ : Union[str, Any] = predictions[: len(eval_dataloader.dataset) - samples_seen] a__ : Any = references[: len(eval_dataloader.dataset) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=_lowercase , references=_lowercase , ) a__ : Optional[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , _lowercase) a__ : Any = eval_metric["""accuracy"""] if best_performance < eval_metric["accuracy"]: a__ : List[str] = eval_metric["""accuracy"""] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), F'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , """all_results.json""") , """w""") as f: json.dump(_lowercase , _lowercase) def lowerCAmelCase_ ( ) -> Tuple: """simple docstring""" a__ : Optional[Any] = argparse.ArgumentParser(description="""Simple example of training script tracking peak GPU memory usage.""") parser.add_argument( """--model_name_or_path""" , type=_lowercase , default="""bert-base-cased""" , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=_lowercase , ) parser.add_argument( """--output_dir""" , type=_lowercase , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--performance_lower_bound""" , type=_lowercase , default=_lowercase , help="""Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.""" , ) parser.add_argument( """--num_epochs""" , type=_lowercase , default=3 , help="""Number of train epochs.""" , ) a__ : Any = parser.parse_args() a__ : Dict = {"""lr""": 2e-5, """num_epochs""": args.num_epochs, """seed""": 42, """batch_size""": 16} training_function(_lowercase , _lowercase) if __name__ == "__main__": main()
266
0
import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def UpperCAmelCase_ ( self : Union[str, Any] , _A : int , _A : int ) -> List[Any]: """simple docstring""" snake_case_ : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def UpperCAmelCase_ ( self : int ) -> List[Any]: """simple docstring""" snake_case_ : Any = None snake_case_ : str = 20 snake_case_ : Optional[int] = self._get_uniform_logits(batch_size=2 , length=_UpperCAmelCase ) # tweak scores to not be uniform anymore snake_case_ : Tuple = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch snake_case_ : Optional[int] = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax snake_case_ : Optional[Any] = jax.nn.softmax(_UpperCAmelCase , axis=-1 ) snake_case_ : Dict = FlaxTemperatureLogitsWarper(temperature=0.5 ) snake_case_ : Union[str, Any] = FlaxTemperatureLogitsWarper(temperature=1.3 ) snake_case_ : List[Any] = jax.nn.softmax(temp_dist_warper_sharper(_UpperCAmelCase , scores.copy() , cur_len=_UpperCAmelCase ) , axis=-1 ) snake_case_ : str = jax.nn.softmax(temp_dist_warper_smoother(_UpperCAmelCase , scores.copy() , cur_len=_UpperCAmelCase ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def UpperCAmelCase_ ( self : List[str] ) -> int: """simple docstring""" snake_case_ : Union[str, Any] = None snake_case_ : Optional[int] = 10 snake_case_ : Tuple = 2 # create ramp distribution snake_case_ : str = np.broadcast_to(np.arange(_UpperCAmelCase )[None, :] , (batch_size, vocab_size) ).copy() snake_case_ : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size snake_case_ : int = FlaxTopKLogitsWarper(3 ) snake_case_ : List[str] = top_k_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case snake_case_ : List[str] = 5 snake_case_ : Optional[int] = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) snake_case_ : Optional[Any] = np.broadcast_to(np.arange(_UpperCAmelCase )[None, :] , (batch_size, length) ).copy() snake_case_ : Union[str, Any] = top_k_warp_safety_check(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" snake_case_ : List[Any] = None snake_case_ : List[Any] = 10 snake_case_ : Any = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) snake_case_ : str = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.1_5, 0.3, 0.3, 0.2_5]] ) ) snake_case_ : Optional[Any] = FlaxTopPLogitsWarper(0.8 ) snake_case_ : List[Any] = np.exp(top_p_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 snake_case_ : int = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.2_5]] ) self.assertTrue(np.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ) ) # check edge cases with negative and extreme logits snake_case_ : Tuple = np.broadcast_to(np.arange(_UpperCAmelCase )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme snake_case_ : Optional[Any] = ramp_logits[1] * 1_0_0.0 # make sure at least 2 tokens are kept snake_case_ : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) snake_case_ : Union[str, Any] = top_p_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def UpperCAmelCase_ ( self : int ) -> Optional[int]: """simple docstring""" snake_case_ : Any = 20 snake_case_ : str = 4 snake_case_ : Tuple = 0 snake_case_ : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_UpperCAmelCase ) # check that min length is applied at length 5 snake_case_ : List[str] = ids_tensor((batch_size, 20) , vocab_size=20 ) snake_case_ : str = 5 snake_case_ : Optional[int] = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : int = min_dist_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('inf' )] ) # check that min length is not applied anymore at length 15 snake_case_ : List[Any] = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Optional[int] = 15 snake_case_ : List[Any] = min_dist_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertFalse(jnp.isinf(_UpperCAmelCase ).any() ) def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" snake_case_ : str = 20 snake_case_ : Dict = 4 snake_case_ : int = 0 snake_case_ : Tuple = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_UpperCAmelCase ) # check that all scores are -inf except the bos_token_id score snake_case_ : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) snake_case_ : Tuple = 1 snake_case_ : Union[str, Any] = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Any = logits_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 snake_case_ : str = 3 snake_case_ : Optional[int] = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Optional[int] = logits_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertFalse(jnp.isinf(_UpperCAmelCase ).any() ) def UpperCAmelCase_ ( self : List[Any] ) -> str: """simple docstring""" snake_case_ : Optional[int] = 20 snake_case_ : List[str] = 4 snake_case_ : Tuple = 0 snake_case_ : Optional[Any] = 5 snake_case_ : Optional[int] = FlaxForcedEOSTokenLogitsProcessor(max_length=_UpperCAmelCase , eos_token_id=_UpperCAmelCase ) # check that all scores are -inf except the eos_token_id when max_length is reached snake_case_ : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) snake_case_ : str = 4 snake_case_ : str = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Dict = logits_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached snake_case_ : int = 3 snake_case_ : str = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : List[Any] = logits_processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) self.assertFalse(jnp.isinf(_UpperCAmelCase ).any() ) def UpperCAmelCase_ ( self : Optional[int] ) -> Tuple: """simple docstring""" snake_case_ : List[str] = 4 snake_case_ : Union[str, Any] = 10 snake_case_ : Optional[Any] = 15 snake_case_ : str = 2 snake_case_ : int = 1 snake_case_ : List[str] = 15 # dummy input_ids and scores snake_case_ : List[Any] = ids_tensor((batch_size, sequence_length) , _UpperCAmelCase ) snake_case_ : List[str] = input_ids.copy() snake_case_ : Any = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : int = scores.copy() # instantiate all dist processors snake_case_ : Dict = FlaxTemperatureLogitsWarper(temperature=0.5 ) snake_case_ : Union[str, Any] = FlaxTopKLogitsWarper(3 ) snake_case_ : Tuple = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors snake_case_ : Optional[int] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_UpperCAmelCase ) snake_case_ : Union[str, Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_UpperCAmelCase ) snake_case_ : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_UpperCAmelCase , eos_token_id=_UpperCAmelCase ) snake_case_ : int = 10 # no processor list snake_case_ : Dict = temp_dist_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Any = top_k_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Union[str, Any] = top_p_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Union[str, Any] = min_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : str = bos_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Dict = eos_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) # with processor list snake_case_ : int = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) snake_case_ : str = processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) # scores should be equal self.assertTrue(jnp.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" snake_case_ : Optional[Any] = 4 snake_case_ : Dict = 10 snake_case_ : Optional[int] = 15 snake_case_ : List[Any] = 2 snake_case_ : Dict = 1 snake_case_ : Tuple = 15 # dummy input_ids and scores snake_case_ : Tuple = ids_tensor((batch_size, sequence_length) , _UpperCAmelCase ) snake_case_ : str = input_ids.copy() snake_case_ : List[str] = self._get_uniform_logits(_UpperCAmelCase , _UpperCAmelCase ) snake_case_ : Union[str, Any] = scores.copy() # instantiate all dist processors snake_case_ : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) snake_case_ : Optional[Any] = FlaxTopKLogitsWarper(3 ) snake_case_ : Dict = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors snake_case_ : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_UpperCAmelCase ) snake_case_ : Tuple = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_UpperCAmelCase ) snake_case_ : List[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=_UpperCAmelCase , eos_token_id=_UpperCAmelCase ) snake_case_ : List[str] = 10 # no processor list def run_no_processor_list(_A : Optional[Any] , _A : Dict , _A : Dict ): snake_case_ : Optional[int] = temp_dist_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Optional[int] = top_k_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Dict = top_p_warp(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Tuple = min_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : int = bos_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) snake_case_ : Dict = eos_dist_proc(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) return scores # with processor list def run_processor_list(_A : Optional[int] , _A : Any , _A : Optional[Any] ): snake_case_ : Tuple = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) snake_case_ : Any = processor(_UpperCAmelCase , _UpperCAmelCase , cur_len=_UpperCAmelCase ) return scores snake_case_ : Union[str, Any] = jax.jit(_UpperCAmelCase ) snake_case_ : str = jax.jit(_UpperCAmelCase ) snake_case_ : int = jitted_run_no_processor_list(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) snake_case_ : List[str] = jitted_run_processor_list(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # scores should be equal self.assertTrue(jnp.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
327
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class A ( tf.keras.layers.Layer ): '''simple docstring''' def __init__(self : Any , _UpperCAmelCase : Dict[str, int] , _UpperCAmelCase : List[str] , _UpperCAmelCase : int = None , _UpperCAmelCase : int = None ) -> Dict: """simple docstring""" super().__init__() lowercase__ = pad_token_id lowercase__ = max_length lowercase__ = vocab lowercase__ = merges lowercase__ = BytePairTokenizer(_UpperCAmelCase , _UpperCAmelCase , sequence_length=_UpperCAmelCase ) @classmethod def lowerCamelCase__ (cls : Optional[int] , _UpperCAmelCase : GPTaTokenizer , *_UpperCAmelCase : List[Any] , **_UpperCAmelCase : List[Any] ) -> Union[str, Any]: """simple docstring""" lowercase__ = [""" """.join(_UpperCAmelCase ) for m in tokenizer.bpe_ranks.keys()] lowercase__ = tokenizer.get_vocab() return cls(_UpperCAmelCase , _UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) @classmethod def lowerCamelCase__ (cls : Union[str, Any] , _UpperCAmelCase : Union[str, os.PathLike] , *_UpperCAmelCase : str , **_UpperCAmelCase : List[Any] ) -> Any: """simple docstring""" lowercase__ = GPTaTokenizer.from_pretrained(_UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) return cls.from_tokenizer(_UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) @classmethod def lowerCamelCase__ (cls : Any , _UpperCAmelCase : Tuple ) -> Union[str, Any]: """simple docstring""" return cls(**_UpperCAmelCase ) def lowerCamelCase__ (self : Union[str, Any] ) -> List[Any]: """simple docstring""" return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def lowerCamelCase__ (self : str , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int = None ) -> Optional[Any]: """simple docstring""" lowercase__ = self.tf_tokenizer(_UpperCAmelCase ) lowercase__ = tf.ones_like(_UpperCAmelCase ) if self.pad_token_id is not None: # pad the tokens up to max length lowercase__ = max_length if max_length is not None else self.max_length if max_length is not None: lowercase__ , lowercase__ = pad_model_inputs( _UpperCAmelCase , max_seq_length=_UpperCAmelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
305
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all MVP models at https://huggingface.co/models?filter=mvp lowerCamelCase__ = { 'vocab_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json', }, 'added_tokens.json': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json', }, 'merges_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt', }, 'tokenizer_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json', }, } lowerCamelCase__ = { 'RUCAIBox/mvp': 1024, } class A__ ( _lowerCamelCase ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ['input_ids', 'attention_mask'] lowercase = MvpTokenizer def __init__( self : Dict , a : List[str]=None , a : Optional[int]=None , a : Union[str, Any]=None , a : Tuple="replace" , a : Dict="<s>" , a : List[str]="</s>" , a : Union[str, Any]="</s>" , a : Optional[Any]="<s>" , a : Optional[int]="<unk>" , a : Optional[Any]="<pad>" , a : int="<mask>" , a : int=False , a : int=True , **a : Any , ): '''simple docstring''' super().__init__( a , a , tokenizer_file=a , errors=a , bos_token=a , eos_token=a , sep_token=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , add_prefix_space=a , trim_offsets=a , **a , ) lowerCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Union[str, Any] = getattr(a , pre_tok_state.pop('type' ) ) lowerCAmelCase__ : Any = add_prefix_space lowerCAmelCase__ : Optional[int] = pre_tok_class(**a ) lowerCAmelCase__ : int = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCAmelCase__ : Optional[Any] = 'post_processor' lowerCAmelCase__ : Optional[int] = getattr(self.backend_tokenizer , a , a ) if tokenizer_component_instance: lowerCAmelCase__ : int = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCAmelCase__ : List[str] = tuple(state['sep'] ) if "cls" in state: lowerCAmelCase__ : str = tuple(state['cls'] ) lowerCAmelCase__ : int = False if state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Optional[int] = add_prefix_space lowerCAmelCase__ : List[Any] = True if state.get('trim_offsets' , a ) != trim_offsets: lowerCAmelCase__ : int = trim_offsets lowerCAmelCase__ : Optional[Any] = True if changes_to_apply: lowerCAmelCase__ : Optional[Any] = getattr(a , state.pop('type' ) ) lowerCAmelCase__ : Tuple = component_class(**a ) setattr(self.backend_tokenizer , a , a ) @property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def _lowerCamelCase ( self : Optional[Any] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : str = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else value lowerCAmelCase__ : Dict = value def _lowerCamelCase ( self : int , *a : Union[str, Any] , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*a , **a ) def _lowerCamelCase ( self : int , *a : Tuple , **a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._encode_plus(*a , **a ) def _lowerCamelCase ( self : List[str] , a : str , a : Optional[str] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self._tokenizer.model.save(a , name=a ) return tuple(a ) def _lowerCamelCase ( self : List[str] , a : List[Any] , a : Optional[int]=None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
353
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCamelCase__ = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""", """BlipTextConfig""", """BlipVisionConfig""", ], """processing_blip""": ["""BlipProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""BlipImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlipModel""", """BlipPreTrainedModel""", """BlipForConditionalGeneration""", """BlipForQuestionAnswering""", """BlipVisionModel""", """BlipTextModel""", """BlipForImageTextRetrieval""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBlipModel""", """TFBlipPreTrainedModel""", """TFBlipForConditionalGeneration""", """TFBlipForQuestionAnswering""", """TFBlipVisionModel""", """TFBlipTextModel""", """TFBlipForImageTextRetrieval""", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
307
0
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCamelCase__ = ['tests.fixtures.files', 'tests.fixtures.hub', 'tests.fixtures.fsspec'] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ["integration", "unit"] ): continue item.add_marker(pytest.mark.unit ) def __lowerCAmelCase (__lowerCAmelCase ): config.addinivalue_line("markers" , "torchaudio_latest: mark test to run with torchaudio>=0.12" ) @pytest.fixture(autouse=lowercase__ ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? _UpperCAmelCase : int = tmp_path_factory.getbasetemp() / '''cache''' _UpperCAmelCase : Optional[Any] = test_hf_cache_home / '''datasets''' _UpperCAmelCase : Tuple = test_hf_cache_home / '''metrics''' _UpperCAmelCase : Optional[int] = test_hf_cache_home / '''modules''' monkeypatch.setattr("datasets.config.HF_DATASETS_CACHE" , str(lowercase__ ) ) monkeypatch.setattr("datasets.config.HF_METRICS_CACHE" , str(lowercase__ ) ) monkeypatch.setattr("datasets.config.HF_MODULES_CACHE" , str(lowercase__ ) ) _UpperCAmelCase : List[str] = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr("datasets.config.DOWNLOADED_DATASETS_PATH" , str(lowercase__ ) ) _UpperCAmelCase : List[str] = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(lowercase__ ) ) @pytest.fixture(autouse=lowercase__ , scope="session" ) def __lowerCAmelCase (): datasets.disable_progress_bar() @pytest.fixture(autouse=lowercase__ ) def __lowerCAmelCase (__lowerCAmelCase ): # don't take tests into account when counting downloads monkeypatch.setattr("datasets.config.HF_UPDATE_DOWNLOAD_COUNTS" , lowercase__ ) @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase ): # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr("sqlalchemy.util.deprecations.SILENCE_UBER_WARNING" , lowercase__ )
234
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = np.full((len(lowercase__ ), sequence_length, 2) , lowercase__ ) else: __lowerCAmelCase : Optional[int] = np.full((len(lowercase__ ), sequence_length) , lowercase__ ) for i, tensor in enumerate(lowercase__ ): if padding_side == "right": if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : int = tensor[:sequence_length] else: if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = ord(lowercase__ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True __lowerCAmelCase : int = unicodedata.category(lowercase__ ) if cat.startswith('''P''' ): return True return False @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = True _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = -100 _UpperCamelCase = "pt" def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCAmelCase : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCAmelCase : List[Any] = self.tokenizer.pad( A_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCAmelCase : Dict = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCAmelCase : Optional[int] = self.tokenizer.padding_side if padding_side == "right": __lowerCAmelCase : Any = [ list(A_ ) + [self.label_pad_token_id] * (sequence_length - len(A_ )) for label in labels ] else: __lowerCAmelCase : Optional[int] = [ [self.label_pad_token_id] * (sequence_length - len(A_ )) + list(A_ ) for label in labels ] __lowerCAmelCase : Tuple = [feature['''ner_tags'''] for feature in features] __lowerCAmelCase : List[Any] = padding_tensor(A_ , -1 , A_ , A_ ) __lowerCAmelCase : Optional[int] = [feature['''original_entity_spans'''] for feature in features] __lowerCAmelCase : Any = padding_tensor(A_ , (-1, -1) , A_ , A_ ) __lowerCAmelCase : Optional[Any] = {k: torch.tensor(A_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
275
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : int = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Dict = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys lowerCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
25
'''simple docstring''' import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase : Tuple = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right lowerCAmelCase : Union[str, Any] = 25_00_04 lowerCAmelCase : int = 25_00_20 @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = MBartaaTokenizer __magic_name__ = MBartaaTokenizerFast __magic_name__ = True __magic_name__ = True def a ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : List[Any] = MBartaaTokenizer(snake_case__ , src_lang='en_XX' , tgt_lang='ro_RO' , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = '<s>' _lowerCAmelCase : str = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(snake_case__ ) , 1054 ) def a ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1054 ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = MBartaaTokenizer(snake_case__ , src_lang='en_XX' , tgt_lang='ro_RO' , keep_accents=snake_case__ ) _lowerCAmelCase : Any = tokenizer.tokenize('This is a test' ) self.assertListEqual(snake_case__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : Tuple = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( snake_case__ , [SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.'] , ) _lowerCAmelCase : Optional[int] = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : Optional[Any] = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.'] , ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = {'input_ids': [[25_0004, 1_1062, 8_2772, 7, 15, 8_2772, 538, 5_1529, 237, 1_7198, 1290, 206, 9, 21_5175, 1314, 136, 1_7198, 1290, 206, 9, 5_6359, 42, 12_2009, 9, 1_6466, 16, 8_7344, 4537, 9, 4717, 7_8381, 6, 15_9958, 7, 15, 2_4480, 618, 4, 527, 2_2693, 5428, 4, 2777, 2_4480, 9874, 4, 4_3523, 594, 4, 803, 1_8392, 3_3189, 18, 4, 4_3523, 2_4447, 1_2399, 100, 2_4955, 8_3658, 9626, 14_4057, 15, 839, 2_2335, 16, 136, 2_4955, 8_3658, 8_3479, 15, 3_9102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 12_2009, 11_5774, 23, 805, 1328, 4_6876, 7, 136, 5_3894, 1940, 4_2227, 4_1159, 1_7721, 823, 425, 4, 2_7512, 9_8722, 206, 136, 5531, 4970, 919, 1_7336, 5, 2], [25_0004, 2_0080, 618, 83, 8_2775, 47, 479, 9, 1517, 73, 5_3894, 333, 8_0581, 11_0117, 1_8811, 5256, 1295, 51, 15_2526, 297, 7986, 390, 12_4416, 538, 3_5431, 214, 98, 1_5044, 2_5737, 136, 7108, 4_3701, 23, 756, 13_5355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_0004, 581, 6_3773, 11_9455, 6, 14_7797, 8_8203, 7, 645, 70, 21, 3285, 1_0269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name='facebook/mbart-large-50' , revision='d3913889c59cd5c9e456b269c376325eabad57e2' , ) def a ( self ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCAmelCase : Optional[int] = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-mbart50', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCAmelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : Tuple = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Tuple = tokenizer_r.save_pretrained(snake_case__ ) _lowerCAmelCase : str = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) _lowerCAmelCase : Any = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : List[str] = tokenizer_r.from_pretrained(snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True _lowerCAmelCase : Union[str, Any] = tempfile.mkdtemp() _lowerCAmelCase : Dict = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) _lowerCAmelCase : Any = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way _lowerCAmelCase : Dict = tokenizer_r.from_pretrained(snake_case__ ) _lowerCAmelCase : List[str] = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False _lowerCAmelCase : Optional[int] = tempfile.mkdtemp() _lowerCAmelCase : int = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) _lowerCAmelCase : Tuple = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCAmelCase : int = tokenizer_r.from_pretrained(snake_case__ ) _lowerCAmelCase : str = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" __magic_name__ = "facebook/mbart-large-50-one-to-many-mmt" __magic_name__ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] __magic_name__ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] __magic_name__ = [EN_CODE, 8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2] @classmethod def a ( cls ): '''simple docstring''' _lowerCAmelCase : MBartaaTokenizer = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en_XX' , tgt_lang='ro_RO' ) _lowerCAmelCase : Dict = 1 return cls def a ( self ): '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ar_AR'] , 25_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['en_EN'] , 25_0004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ro_RO'] , 25_0020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['mr_IN'] , 25_0038 ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , snake_case__ ) def a ( self ): '''simple docstring''' self.assertIn(snake_case__ , self.tokenizer.all_special_ids ) _lowerCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 8_6792, 36, 1_8743, 1_5596, 5, 2] _lowerCAmelCase : List[str] = self.tokenizer.decode(snake_case__ , skip_special_tokens=snake_case__ ) _lowerCAmelCase : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=snake_case__ ) self.assertEqual(snake_case__ , snake_case__ ) self.assertNotIn(self.tokenizer.eos_token , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , snake_case__ ) _lowerCAmelCase : List[str] = 10 _lowerCAmelCase : Any = self.tokenizer(snake_case__ , max_length=snake_case__ , truncation=snake_case__ ).input_ids[0] self.assertEqual(ids[0] , snake_case__ ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(snake_case__ ) , snake_case__ ) def a ( self ): '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', 'ar_AR'] ) , [25_0053, 25_0001] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = tempfile.mkdtemp() _lowerCAmelCase : Dict = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(snake_case__ ) _lowerCAmelCase : Tuple = MBartaaTokenizer.from_pretrained(snake_case__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , snake_case__ ) @require_torch def a ( self ): '''simple docstring''' _lowerCAmelCase : str = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=snake_case__ , return_tensors='pt' ) _lowerCAmelCase : Optional[int] = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def a ( self ): '''simple docstring''' _lowerCAmelCase : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=snake_case__ , truncation=snake_case__ , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) _lowerCAmelCase : int = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(snake_case__ , snake_case__ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) _lowerCAmelCase : Union[str, Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , snake_case__ ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , padding=snake_case__ , truncation=snake_case__ , max_length=3 , return_tensors='pt' ) _lowerCAmelCase : str = self.tokenizer( text_target=self.tgt_text , padding=snake_case__ , truncation=snake_case__ , max_length=10 , return_tensors='pt' ) _lowerCAmelCase : List[Any] = targets['input_ids'] _lowerCAmelCase : Any = shift_tokens_right(snake_case__ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='ar_AR' ) self.assertEqual( nested_simplify(snake_case__ ) , { # en_XX, A, test, EOS 'input_ids': [[25_0004, 62, 3034, 2]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 25_0001, } , )
25
1
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : Union[str, Any]) -> Optional[Any]: __snake_case : int = 0 def _lowercase (self : Union[str, Any]) -> Optional[Any]: __snake_case : int = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32') self.assertIsInstance(_A , _A) def _lowercase (self : Any) -> Tuple: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Union[str, Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : List[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Any = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[str]) -> Optional[int]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = CLIPConfig() # Create a dummy config file with image_proceesor_type __snake_case : int = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __snake_case : List[Any] = AutoImageProcessor.from_pretrained(_A).to_dict() config_dict.pop('image_processor_type') __snake_case : Dict = CLIPImageProcessor(**_A) # save in new folder model_config.save_pretrained(_A) config.save_pretrained(_A) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) # make sure private variable is not incorrectly saved __snake_case : Optional[int] = json.loads(config.to_json_string()) self.assertTrue('_processor_class' not in dict_as_saved) self.assertIsInstance(_A , _A) def _lowercase (self : Union[str, Any]) -> str: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = Path(_A) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) __snake_case : Any = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> str: with self.assertRaisesRegex( _A , 'clip-base is not a local folder and is not a valid model identifier'): __snake_case : Tuple = AutoImageProcessor.from_pretrained('clip-base') def _lowercase (self : Optional[int]) -> int: with self.assertRaisesRegex( _A , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'): __snake_case : int = AutoImageProcessor.from_pretrained(_A , revision='aaaaaa') def _lowercase (self : Tuple) -> Optional[int]: with self.assertRaisesRegex( _A , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __snake_case : Any = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model') def _lowercase (self : Tuple) -> Tuple: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_A): __snake_case : Optional[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') # If remote code is disabled, we can't load this config. with self.assertRaises(_A): __snake_case : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) __snake_case : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Dict = AutoImageProcessor.from_pretrained(_A , trust_remote_code=_A) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor') def _lowercase (self : int) -> str: try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A): AutoImageProcessor.register(_A , _A) with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Optional[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = CustomImageProcessor.from_pretrained(_A) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowercase (self : int) -> str: class UpperCamelCase ( lowercase ): UpperCAmelCase : List[str] = True try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # If remote code is not set, the default is to use local __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote code is disabled, we load the local one. __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote is enabled, we load from the Hub __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(not hasattr(_A , 'is_local')) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
172
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Union[str, Any]= logging.get_logger(__name__) _a : str= { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCamelCase ( lowercase ): UpperCAmelCase : List[str] = """mgp-str""" def __init__(self : List[Any] , _A : Dict=[32, 1_28] , _A : Any=4 , _A : int=3 , _A : Any=27 , _A : List[str]=38 , _A : str=5_02_57 , _A : Optional[int]=3_05_22 , _A : Union[str, Any]=7_68 , _A : Tuple=12 , _A : List[str]=12 , _A : List[str]=4.0 , _A : Optional[int]=True , _A : Optional[Any]=False , _A : Dict=1E-5 , _A : Optional[int]=0.0 , _A : str=0.0 , _A : int=0.0 , _A : str=False , _A : List[Any]=0.02 , **_A : Union[str, Any] , ) -> Tuple: super().__init__(**_A) __snake_case : Union[str, Any] = image_size __snake_case : Optional[int] = patch_size __snake_case : int = num_channels __snake_case : int = max_token_length __snake_case : List[Any] = num_character_labels __snake_case : Optional[int] = num_bpe_labels __snake_case : Optional[Any] = num_wordpiece_labels __snake_case : int = hidden_size __snake_case : List[Any] = num_hidden_layers __snake_case : List[str] = num_attention_heads __snake_case : Any = mlp_ratio __snake_case : List[str] = distilled __snake_case : List[Any] = layer_norm_eps __snake_case : List[Any] = drop_rate __snake_case : Optional[int] = qkv_bias __snake_case : Optional[int] = attn_drop_rate __snake_case : int = drop_path_rate __snake_case : List[str] = output_aa_attentions __snake_case : Optional[Any] = initializer_range
172
1
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : Tuple = { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json", } class A( __lowercase ): '''simple docstring''' UpperCamelCase = "mvp" UpperCamelCase = ["past_key_values"] UpperCamelCase = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[int] , A_ : List[str]=50267 , A_ : Dict=1024 , A_ : Optional[Any]=12 , A_ : Tuple=4096 , A_ : Dict=16 , A_ : Tuple=12 , A_ : int=4096 , A_ : List[Any]=16 , A_ : str=0.0 , A_ : List[Any]=0.0 , A_ : List[Any]="gelu" , A_ : Dict=1024 , A_ : List[str]=0.1 , A_ : Tuple=0.0 , A_ : Optional[int]=0.0 , A_ : Optional[Any]=0.02 , A_ : Optional[int]=0.0 , A_ : str=False , A_ : List[str]=True , A_ : Any=1 , A_ : List[str]=0 , A_ : Dict=2 , A_ : Any=True , A_ : List[Any]=2 , A_ : str=2 , A_ : str=False , A_ : List[Any]=100 , A_ : int=800 , **A_ : Optional[int] , ) -> List[str]: """simple docstring""" lowerCamelCase_ = vocab_size lowerCamelCase_ = max_position_embeddings lowerCamelCase_ = d_model lowerCamelCase_ = encoder_ffn_dim lowerCamelCase_ = encoder_layers lowerCamelCase_ = encoder_attention_heads lowerCamelCase_ = decoder_ffn_dim lowerCamelCase_ = decoder_layers lowerCamelCase_ = decoder_attention_heads lowerCamelCase_ = dropout lowerCamelCase_ = attention_dropout lowerCamelCase_ = activation_dropout lowerCamelCase_ = activation_function lowerCamelCase_ = init_std lowerCamelCase_ = encoder_layerdrop lowerCamelCase_ = decoder_layerdrop lowerCamelCase_ = classifier_dropout lowerCamelCase_ = use_cache lowerCamelCase_ = encoder_layers lowerCamelCase_ = scale_embedding # scale factor will be sqrt(d_model) if True lowerCamelCase_ = use_prompt lowerCamelCase_ = prompt_length lowerCamelCase_ = prompt_mid_dim super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , is_encoder_decoder=_a , decoder_start_token_id=_a , forced_eos_token_id=_a , **_a , ) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , _a ): lowerCamelCase_ = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ 'The config can simply be saved and uploaded again to be fixed.' )
365
from ..utils import DummyObject, requires_backends class A( metaclass=UpperCamelCase ): '''simple docstring''' UpperCamelCase = ['''transformers''', '''torch''', '''note_seq'''] def __init__( self : Any , *A_ : Any , **A_ : Tuple ) -> List[Any]: """simple docstring""" requires_backends(self , ['transformers', 'torch', 'note_seq'] ) @classmethod def a__ ( cls : Tuple , *A_ : Dict , **A_ : Optional[Any] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ['transformers', 'torch', 'note_seq'] ) @classmethod def a__ ( cls : int , *A_ : Dict , **A_ : List[str] ) -> Tuple: """simple docstring""" requires_backends(cls , ['transformers', 'torch', 'note_seq'] )
208
0
from collections.abc import Iterable from typing import Any class A__ : def __init__( self : Union[str, Any] , _UpperCAmelCase : Tuple = None ) -> Optional[Any]: """simple docstring""" __lowercase = value __lowercase = None # Added in order to delete a node easier __lowercase = None __lowercase = None def __repr__( self : str ) -> str: """simple docstring""" from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({f"""{self.value}""": (self.left, self.right)} , indent=1 ) class A__ : def __init__( self : int , _UpperCAmelCase : Optional[Any] = None ) -> str: """simple docstring""" __lowercase = root def __str__( self : List[Any] ) -> str: """simple docstring""" return str(self.root ) def a__ ( self : int , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ) -> None: """simple docstring""" if new_children is not None: # reset its kids __lowercase = node.parent if node.parent is not None: # reset its parent if self.is_right(lowercase_ ): # If it is the right children __lowercase = new_children else: __lowercase = new_children else: __lowercase = new_children def a__ ( self : Tuple , _UpperCAmelCase : Any ) -> bool: """simple docstring""" if node.parent and node.parent.right: return node == node.parent.right return False def a__ ( self : List[str] ) -> bool: """simple docstring""" return self.root is None def a__ ( self : Tuple , _UpperCAmelCase : Optional[int] ) -> None: """simple docstring""" __lowercase = Node(lowercase_ ) # create a new Node if self.empty(): # if Tree is empty __lowercase = new_node # set its root else: # Tree is not empty __lowercase = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: __lowercase = new_node # We insert the new node in a leaf break else: __lowercase = parent_node.left else: if parent_node.right is None: __lowercase = new_node break else: __lowercase = parent_node.right __lowercase = parent_node def a__ ( self : List[str] , *_UpperCAmelCase : Union[str, Any] ) -> None: """simple docstring""" for value in values: self.__insert(lowercase_ ) def a__ ( self : List[Any] , _UpperCAmelCase : Optional[int] ) -> Node | None: """simple docstring""" if self.empty(): raise IndexError('Warning: Tree is empty! please use another.' ) else: __lowercase = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: __lowercase = node.left if value < node.value else node.right return node def a__ ( self : Dict , _UpperCAmelCase : List[Any] = None ) -> Node | None: """simple docstring""" if node is None: if self.root is None: return None __lowercase = self.root if not self.empty(): while node.right is not None: __lowercase = node.right return node def a__ ( self : int , _UpperCAmelCase : List[Any] = None ) -> Node | None: """simple docstring""" if node is None: __lowercase = self.root if self.root is None: return None if not self.empty(): __lowercase = self.root while node.left is not None: __lowercase = node.left return node def a__ ( self : Optional[int] , _UpperCAmelCase : Tuple ) -> None: """simple docstring""" __lowercase = self.search(lowercase_ ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(lowercase_ , lowercase_ ) elif node.left is None: # Has only right children self.__reassign_nodes(lowercase_ , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(lowercase_ , node.left ) else: __lowercase = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore __lowercase = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def a__ ( self : List[Any] , _UpperCAmelCase : str ) -> Iterable: """simple docstring""" if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def a__ ( self : List[Any] , _UpperCAmelCase : Optional[Any]=None ) -> Any: """simple docstring""" if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def a__ ( self : List[str] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> None: """simple docstring""" if node: self.inorder(lowercase_ , node.left ) arr.append(node.value ) self.inorder(lowercase_ , node.right ) def a__ ( self : List[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : int ) -> int: """simple docstring""" __lowercase = [] self.inorder(lowercase_ , lowercase_ ) # append all values to list using inorder traversal return arr[k - 1] def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE : Node | None ) -> Dict: __lowercase = [] if curr_node is not None: __lowercase = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def __SCREAMING_SNAKE_CASE ( ) -> List[Any]: __lowercase = (8, 3, 6, 1, 10, 14, 13, 4, 7) __lowercase = BinarySearchTree() for i in testlist: t.insert(_A ) # Prints all the elements of the list in order traversal print(_A ) if t.search(6 ) is not None: print('The value 6 exists' ) else: print('The value 6 doesn\'t exist' ) if t.search(-1 ) is not None: print('The value -1 exists' ) else: print('The value -1 doesn\'t exist' ) if not t.empty(): print('Max Value: ' , t.get_max().value ) # type: ignore print('Min Value: ' , t.get_min().value ) # type: ignore for i in testlist: t.remove(_A ) print(_A ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
325
from __future__ import annotations def UpperCAmelCase__ ( _A : float , _A : float , _A : float , ): '''simple docstring''' if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
188
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
370
"""simple docstring""" def __UpperCAmelCase ( lowercase = 10_00 ): """simple docstring""" _UpperCAmelCase = 2**power _UpperCAmelCase = 0 while n: _UpperCAmelCase , _UpperCAmelCase = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
30
0
from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCamelCase : def __init__( self :int , lowerCamelCase :Tuple , lowerCamelCase :Optional[int]=13 , lowerCamelCase :Union[str, Any]=30 , lowerCamelCase :List[Any]=2 , lowerCamelCase :Tuple=3 , lowerCamelCase :Tuple=True , lowerCamelCase :Optional[int]=True , lowerCamelCase :Optional[Any]=32 , lowerCamelCase :Optional[Any]=2 , lowerCamelCase :List[Any]=4 , lowerCamelCase :Tuple=37 , lowerCamelCase :List[Any]="gelu" , lowerCamelCase :List[str]=0.1 , lowerCamelCase :List[str]=0.1 , lowerCamelCase :int=10 , lowerCamelCase :Tuple=0.02 , lowerCamelCase :Optional[int]=3 , lowerCamelCase :Dict=0.6 , lowerCamelCase :int=None , ) -> List[str]: UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = patch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = mask_ratio UpperCAmelCase__ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) UpperCAmelCase__ = (image_size // patch_size) ** 2 UpperCAmelCase__ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def UpperCAmelCase_ ( self :Tuple ) -> Optional[Any]: UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self :str ) -> List[str]: return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase_ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def UpperCAmelCase_ ( self :Optional[int] , lowerCamelCase :Union[str, Any] , lowerCamelCase :Dict , lowerCamelCase :Optional[Any] ) -> List[Any]: UpperCAmelCase__ = TFViTMAEModel(config=UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , training=UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self :Optional[int] , lowerCamelCase :int , lowerCamelCase :str , lowerCamelCase :int ) -> Any: UpperCAmelCase__ = TFViTMAEForPreTraining(UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , training=UpperCamelCase_ ) # expected sequence length = num_patches UpperCAmelCase__ = (self.image_size // self.patch_size) ** 2 UpperCAmelCase__ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images UpperCAmelCase__ = 1 UpperCAmelCase__ = TFViTMAEForPreTraining(UpperCamelCase_ ) UpperCAmelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ = model(UpperCamelCase_ , training=UpperCamelCase_ ) UpperCAmelCase__ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def UpperCAmelCase_ ( self :Tuple ) -> Union[str, Any]: UpperCAmelCase__ = self.prepare_config_and_inputs() ((UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__)) = config_and_inputs UpperCAmelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): UpperCAmelCase_ = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () UpperCAmelCase_ = {'''feature-extraction''': TFViTMAEModel} if is_tf_available() else {} UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = False UpperCAmelCase_ = False def UpperCAmelCase_ ( self :List[str] ) -> Union[str, Any]: UpperCAmelCase__ = TFViTMAEModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_ , hidden_size=37 ) def UpperCAmelCase_ ( self :Union[str, Any] ) -> str: self.config_tester.run_common_tests() @unittest.skip(reason="ViTMAE does not use inputs_embeds" ) def UpperCAmelCase_ ( self :int ) -> Tuple: pass def UpperCAmelCase_ ( self :int ) -> Tuple: UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase_ , tf.keras.layers.Layer ) ) def UpperCAmelCase_ ( self :Union[str, Any] ) -> Tuple: UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) UpperCAmelCase__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_ ) def UpperCAmelCase_ ( self :List[Any] ) -> List[str]: UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_ ) def UpperCAmelCase_ ( self :Tuple ) -> Any: UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase_ ) def UpperCAmelCase_ ( self :Union[str, Any] ) -> str: np.random.seed(2 ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = int((config.image_size // config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) UpperCAmelCase__ = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , noise=UpperCamelCase_ ) UpperCAmelCase__ = copy.deepcopy(self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) ) UpperCAmelCase__ = model(**UpperCamelCase_ , noise=UpperCamelCase_ ) UpperCAmelCase__ = outputs_dict[0].numpy() UpperCAmelCase__ = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def UpperCAmelCase_ ( self :Dict ) -> Tuple: np.random.seed(2 ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = int((config.image_size // config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase :Union[str, Any] ): UpperCAmelCase__ = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase_ ): UpperCAmelCase__ = v.numpy() else: UpperCAmelCase__ = np.array(UpperCamelCase_ ) return inputs_np_dict for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) UpperCAmelCase__ = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ = prepare_numpy_arrays(UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , noise=UpperCamelCase_ ) UpperCAmelCase__ = model(**UpperCamelCase_ , noise=UpperCamelCase_ ) self.assert_outputs_same(UpperCamelCase_ , UpperCamelCase_ ) def UpperCAmelCase_ ( self :str , lowerCamelCase :Tuple , lowerCamelCase :List[str] , lowerCamelCase :List[Any] ) -> Union[str, Any]: np.random.seed(2 ) UpperCAmelCase__ = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCAmelCase__ = tf.constant(UpperCamelCase_ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument UpperCAmelCase__ = tf_noise super().check_pt_tf_models(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) def UpperCAmelCase_ ( self :str ) -> int: np.random.seed(2 ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase_ ) if module_member_name.endswith("MainLayer" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("MainLayer" )] == model_class.__name__[: -len("Model" )] for module_member in (getattr(UpperCamelCase_ , UpperCamelCase_ ),) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase_ , "_keras_serializable" , UpperCamelCase_ ) } UpperCAmelCase__ = int((config.image_size // config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCAmelCase__ = tf.convert_to_tensor(UpperCamelCase_ ) inputs_dict.update({"noise": noise} ) for main_layer_class in tf_main_layer_classes: UpperCAmelCase__ = main_layer_class(UpperCamelCase_ ) UpperCAmelCase__ = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } UpperCAmelCase__ = tf.keras.Model(UpperCamelCase_ , outputs=main_layer(UpperCamelCase_ ) ) UpperCAmelCase__ = model(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = os.path.join(UpperCamelCase_ , "keras_model.h5" ) model.save(UpperCamelCase_ ) UpperCAmelCase__ = tf.keras.models.load_model( UpperCamelCase_ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase_ , tf.keras.Model ) UpperCAmelCase__ = model(UpperCamelCase_ ) self.assert_outputs_same(UpperCamelCase_ , UpperCamelCase_ ) @slow def UpperCAmelCase_ ( self :Optional[int] ) -> Tuple: np.random.seed(2 ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = int((config.image_size // config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) UpperCAmelCase__ = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , noise=UpperCamelCase_ ) if model_class.__name__ == "TFViTMAEModel": UpperCAmelCase__ = outputs.last_hidden_state.numpy() UpperCAmelCase__ = 0 else: UpperCAmelCase__ = outputs.logits.numpy() UpperCAmelCase__ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase_ , saved_model=UpperCamelCase_ ) UpperCAmelCase__ = model_class.from_pretrained(UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , noise=UpperCamelCase_ ) if model_class.__name__ == "TFViTMAEModel": UpperCAmelCase__ = after_outputs["last_hidden_state"].numpy() UpperCAmelCase__ = 0 else: UpperCAmelCase__ = after_outputs["logits"].numpy() UpperCAmelCase__ = 0 UpperCAmelCase__ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase_ , 1e-5 ) def UpperCAmelCase_ ( self :Union[str, Any] ) -> int: np.random.seed(2 ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = int((config.image_size // config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(UpperCamelCase_ ) UpperCAmelCase__ = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ = model(UpperCamelCase_ , noise=UpperCamelCase_ ) UpperCAmelCase__ = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase_ ) UpperCAmelCase__ = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config UpperCAmelCase__ = model_class.from_config(model.config ) UpperCAmelCase__ = new_model(UpperCamelCase_ ) # Build model new_model.set_weights(model.get_weights() ) UpperCAmelCase__ = new_model(UpperCamelCase_ , noise=UpperCamelCase_ ) self.assert_outputs_same(UpperCamelCase_ , UpperCamelCase_ ) @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def UpperCAmelCase_ ( self :Tuple ) -> str: pass @unittest.skip(reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load" ) def UpperCAmelCase_ ( self :Union[str, Any] ) -> List[str]: pass @slow def UpperCAmelCase_ ( self :str ) -> List[str]: UpperCAmelCase__ = TFViTMAEModel.from_pretrained("google/vit-base-patch16-224" ) self.assertIsNotNone(UpperCamelCase_ ) def lowerCAmelCase ( ): """simple docstring""" UpperCAmelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class _UpperCamelCase ( unittest.TestCase ): @cached_property def UpperCAmelCase_ ( self :List[str] ) -> List[str]: return ViTImageProcessor.from_pretrained("facebook/vit-mae-base" ) if is_vision_available() else None @slow def UpperCAmelCase_ ( self :int ) -> Optional[Any]: np.random.seed(2 ) UpperCAmelCase__ = TFViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base" ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=UpperCamelCase_ , return_tensors="tf" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) UpperCAmelCase__ = ViTMAEConfig() UpperCAmelCase__ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) UpperCAmelCase__ = np.random.uniform(size=(1, num_patches) ) # forward pass UpperCAmelCase__ = model(**UpperCamelCase_ , noise=UpperCamelCase_ ) # verify the logits UpperCAmelCase__ = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase_ ) UpperCAmelCase__ = tf.convert_to_tensor( [[-0.05_48, -1.70_23, -0.93_25], [0.37_21, -0.56_70, -0.22_33], [0.82_35, -1.38_78, -0.35_24]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase_ , atol=1e-4 )
169
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging lowerCAmelCase = logging.get_logger(__name__) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: with open(SCREAMING_SNAKE_CASE , '''rb''' ) as flax_state_f: lowercase__ = from_bytes(SCREAMING_SNAKE_CASE , flax_state_f.read() ) except UnpicklingError as e: try: with open(SCREAMING_SNAKE_CASE ) as f: if f.read().startswith('''version''' ): raise OSError( '''You seem to have cloned a repository without having git-lfs installed. Please''' ''' install git-lfs and run `git lfs install` followed by `git lfs pull` in the''' ''' folder you cloned.''' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f'Unable to convert {model_file} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( '''Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights lowercase__ = flatten_dict(jax.tree_util.tree_map(lambda SCREAMING_SNAKE_CASE : x.dtype == jnp.bfloataa , SCREAMING_SNAKE_CASE ) ).values() if any(SCREAMING_SNAKE_CASE ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) lowercase__ = jax.tree_util.tree_map( lambda SCREAMING_SNAKE_CASE : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , SCREAMING_SNAKE_CASE ) lowercase__ = '''''' lowercase__ = flatten_dict(SCREAMING_SNAKE_CASE , sep='''.''' ) lowercase__ = pt_model.state_dict() # keep track of unexpected & missing keys lowercase__ = [] lowercase__ = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowercase__ = flax_key_tuple.split('''.''' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = jnp.transpose(SCREAMING_SNAKE_CASE , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(SCREAMING_SNAKE_CASE ): lowercase__ = ( flax_key_tuple_string.replace('''_0''' , '''.0''' ) .replace('''_1''' , '''.1''' ) .replace('''_2''' , '''.2''' ) .replace('''_3''' , '''.3''' ) .replace('''_4''' , '''.4''' ) .replace('''_5''' , '''.5''' ) .replace('''_6''' , '''.6''' ) .replace('''_7''' , '''.7''' ) .replace('''_8''' , '''.8''' ) .replace('''_9''' , '''.9''' ) ) lowercase__ = '''.'''.join(SCREAMING_SNAKE_CASE ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' f'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict lowercase__ = np.asarray(SCREAMING_SNAKE_CASE ) if not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) else flax_tensor lowercase__ = torch.from_numpy(SCREAMING_SNAKE_CASE ) # remove from missing keys missing_keys.remove(SCREAMING_SNAKE_CASE ) else: # weight is not expected by PyTorch model unexpected_keys.append(SCREAMING_SNAKE_CASE ) pt_model.load_state_dict(SCREAMING_SNAKE_CASE ) # re-transform missing_keys to list lowercase__ = list(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' f' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' f' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' f' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( f'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' f' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ''' use it for predictions and inference.''' ) return pt_model
110
0
"""simple docstring""" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = 0 __lowerCAmelCase = len(_UpperCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(_UpperCamelCase ): return None __lowerCAmelCase = sorted_collection[point] if current_item == item: return point else: if point < left: __lowerCAmelCase = left __lowerCAmelCase = point elif point > right: __lowerCAmelCase = right __lowerCAmelCase = point else: if item < current_item: __lowerCAmelCase = point - 1 else: __lowerCAmelCase = point + 1 return None def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(_UpperCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) elif point > right: return interpolation_search_by_recursion(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , point - 1 ) else: return interpolation_search_by_recursion( _UpperCamelCase , _UpperCamelCase , point + 1 , _UpperCamelCase ) def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' if collection != sorted(_UpperCamelCase ): raise ValueError("Collection must be ascending sorted" ) return True if __name__ == "__main__": import sys A : Any = 0 if debug == 1: A : int = [1_0, 3_0, 4_0, 4_5, 5_0, 6_6, 7_7, 9_3] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") A : Union[str, Any] = 6_7 A : Tuple = interpolation_search(collection, target) if result is not None: print(f'''{target} found at positions: {result}''') else: print("Not found")
259
"""simple docstring""" import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset A : Any = "bert-base-cased" A : Any = "google/pegasus-xsum" A : Union[str, Any] = [" Sam ate lunch today.", "Sams lunch ingredients."] A : Union[str, Any] = ["A very interesting story about what I ate for lunch.", "Avocado, celery, turkey, coffee"] A : Optional[int] = "patrickvonplaten/t5-tiny-random" A : int = "sshleifer/bart-tiny-random" A : Optional[int] = "sshleifer/tiny-mbart" A : Any = "sshleifer/tiny-marian-en-de" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = "\n".join(_UpperCamelCase ) Path(_UpperCamelCase ).open("w" ).writelines(_UpperCamelCase ) def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' for split in ["train", "val", "test"]: _dump_articles(os.path.join(_UpperCamelCase , f"{split}.source" ) , _UpperCamelCase ) _dump_articles(os.path.join(_UpperCamelCase , f"{split}.target" ) , _UpperCamelCase ) return tmp_dir class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def snake_case ( self , __a ): __lowerCAmelCase = AutoTokenizer.from_pretrained(__a ) __lowerCAmelCase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __lowerCAmelCase = max(len(tokenizer.encode(__a ) ) for a in ARTICLES ) __lowerCAmelCase = max(len(tokenizer.encode(__a ) ) for a in SUMMARIES ) __lowerCAmelCase = 4 __lowerCAmelCase = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated __lowerCAmelCase , __lowerCAmelCase = "ro_RO", "de_DE" # ignored for all but mbart, but never causes error. __lowerCAmelCase = SeqaSeqDataset( __a , data_dir=__a , type_path="train" , max_source_length=__a , max_target_length=__a , src_lang=__a , tgt_lang=__a , ) __lowerCAmelCase = DataLoader(__a , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(__a , __a ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place __lowerCAmelCase = shift_tokens_right(batch["labels"] , tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def snake_case ( self , __a ): __lowerCAmelCase = AutoTokenizer.from_pretrained(__a ) __lowerCAmelCase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __lowerCAmelCase = max(len(tokenizer.encode(__a ) ) for a in ARTICLES ) __lowerCAmelCase = max(len(tokenizer.encode(__a ) ) for a in SUMMARIES ) __lowerCAmelCase = 4 __lowerCAmelCase = LegacySeqaSeqDataset( __a , data_dir=__a , type_path="train" , max_source_length=20 , max_target_length=__a , ) __lowerCAmelCase = DataLoader(__a , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 20 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def snake_case ( self ): __lowerCAmelCase = AutoTokenizer.from_pretrained("facebook/mbart-large-cc25" ) __lowerCAmelCase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) __lowerCAmelCase = tmp_dir.joinpath("train.source" ).open().readlines() __lowerCAmelCase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(__a , __a , 1_28 , __a ) __lowerCAmelCase = {x.name for x in tmp_dir.iterdir()} __lowerCAmelCase = {x.name for x in save_dir.iterdir()} __lowerCAmelCase = save_dir.joinpath("train.source" ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(__a ) < len(__a ) assert len(__a ) == 1 assert len(packed_examples[0] ) == sum(len(__a ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason="This test requires fairseq" ) def snake_case ( self ): if not FAIRSEQ_AVAILABLE: return __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self._get_dataset(max_len=64 ) __lowerCAmelCase = 64 __lowerCAmelCase = ds.make_dynamic_sampler(__a , required_batch_size_multiple=__a ) __lowerCAmelCase = [len(__a ) for x in batch_sampler] assert len(set(__a ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(__a ) == len(__a ) # no dropped or added examples __lowerCAmelCase = DataLoader(__a , batch_sampler=__a , collate_fn=ds.collate_fn , num_workers=2 ) __lowerCAmelCase = [] __lowerCAmelCase = [] for batch in data_loader: __lowerCAmelCase = batch["input_ids"].shape __lowerCAmelCase = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple __lowerCAmelCase = np.product(batch["input_ids"].shape ) num_src_per_batch.append(__a ) if num_src_tokens > (max_tokens * 1.1): failures.append(__a ) assert num_src_per_batch[0] == max(__a ) if failures: raise AssertionError(f"too many tokens in {len(__a )} batches" ) def snake_case ( self ): __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self._get_dataset(max_len=5_12 ) __lowerCAmelCase = 2 __lowerCAmelCase = ds.make_sortish_sampler(__a , shuffle=__a ) __lowerCAmelCase = DataLoader(__a , batch_size=__a , collate_fn=ds.collate_fn , num_workers=2 ) __lowerCAmelCase = DataLoader(__a , batch_size=__a , collate_fn=ds.collate_fn , num_workers=2 , sampler=__a ) __lowerCAmelCase = tokenizer.pad_token_id def count_pad_tokens(__a , __a="input_ids" ): return [batch[k].eq(__a ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(__a , k="labels" ) ) < sum(count_pad_tokens(__a , k="labels" ) ) assert sum(count_pad_tokens(__a ) ) < sum(count_pad_tokens(__a ) ) assert len(__a ) == len(__a ) def snake_case ( self , __a=10_00 , __a=1_28 ): if os.getenv("USE_REAL_DATA" , __a ): __lowerCAmelCase = "examples/seq2seq/wmt_en_ro" __lowerCAmelCase = max_len * 2 * 64 if not Path(__a ).joinpath("train.len" ).exists(): save_len_file(__a , __a ) else: __lowerCAmelCase = "examples/seq2seq/test_data/wmt_en_ro" __lowerCAmelCase = max_len * 4 save_len_file(__a , __a ) __lowerCAmelCase = AutoTokenizer.from_pretrained(__a ) __lowerCAmelCase = SeqaSeqDataset( __a , data_dir=__a , type_path="train" , max_source_length=__a , max_target_length=__a , n_obs=__a , ) return ds, max_tokens, tokenizer def snake_case ( self ): __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self._get_dataset() __lowerCAmelCase = set(DistributedSortishSampler(__a , 2_56 , num_replicas=2 , rank=0 , add_extra_examples=__a ) ) __lowerCAmelCase = set(DistributedSortishSampler(__a , 2_56 , num_replicas=2 , rank=1 , add_extra_examples=__a ) ) assert idsa.intersection(__a ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def snake_case ( self , __a ): __lowerCAmelCase = AutoTokenizer.from_pretrained(__a , use_fast=__a ) if tok_name == MBART_TINY: __lowerCAmelCase = SeqaSeqDataset( __a , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path="train" , max_source_length=4 , max_target_length=8 , src_lang="EN" , tgt_lang="FR" , ) __lowerCAmelCase = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: __lowerCAmelCase = SeqaSeqDataset( __a , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path="train" , max_source_length=4 , max_target_length=8 , ) __lowerCAmelCase = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(__a ) == 1 if tok_name == BART_TINY else len(__a ) == 0
259
1
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __lowerCAmelCase : @staticmethod def UpperCAmelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_timm @require_torch class __lowerCAmelCase ( unittest.TestCase ): lowercase = MODEL_FOR_OBJECT_DETECTION_MAPPING def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = ObjectDetectionPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = object_detector('./tests/fixtures/tests_samples/COCO/000000039769.png' , threshold=0.0 ) self.assertGreater(len(__UpperCAmelCase ) , 0 ) for detected_object in outputs: self.assertEqual( __UpperCAmelCase , { 'score': ANY(__UpperCAmelCase ), 'label': ANY(__UpperCAmelCase ), 'box': {'xmin': ANY(__UpperCAmelCase ), 'ymin': ANY(__UpperCAmelCase ), 'xmax': ANY(__UpperCAmelCase ), 'ymax': ANY(__UpperCAmelCase )}, } , ) import datasets __UpperCamelCase = datasets.load_dataset('hf-internal-testing/fixtures_image_utils' , 'image' , split='test' ) __UpperCamelCase = [ Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ), 'http://images.cocodataset.org/val2017/000000039769.jpg', # RGBA dataset[0]['file'], # LA dataset[1]['file'], # L dataset[2]['file'], ] __UpperCamelCase = object_detector(__UpperCAmelCase , threshold=0.0 ) self.assertEqual(len(__UpperCAmelCase ) , len(__UpperCAmelCase ) ) for outputs in batch_outputs: self.assertGreater(len(__UpperCAmelCase ) , 0 ) for detected_object in outputs: self.assertEqual( __UpperCAmelCase , { 'score': ANY(__UpperCAmelCase ), 'label': ANY(__UpperCAmelCase ), 'box': {'xmin': ANY(__UpperCAmelCase ), 'ymin': ANY(__UpperCAmelCase ), 'xmax': ANY(__UpperCAmelCase ), 'ymax': ANY(__UpperCAmelCase )}, } , ) @require_tf @unittest.skip('Object detection not implemented in TF' ) def UpperCAmelCase ( self ): '''simple docstring''' pass @require_torch def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'hf-internal-testing/tiny-detr-mobilenetsv3' __UpperCamelCase = AutoModelForObjectDetection.from_pretrained(__UpperCAmelCase ) __UpperCamelCase = AutoFeatureExtractor.from_pretrained(__UpperCAmelCase ) __UpperCamelCase = ObjectDetectionPipeline(model=__UpperCAmelCase , feature_extractor=__UpperCAmelCase ) __UpperCamelCase = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=0.0 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ] , ) __UpperCamelCase = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], ] , ) @require_torch @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'facebook/detr-resnet-50' __UpperCamelCase = AutoModelForObjectDetection.from_pretrained(__UpperCAmelCase ) __UpperCamelCase = AutoFeatureExtractor.from_pretrained(__UpperCAmelCase ) __UpperCamelCase = ObjectDetectionPipeline(model=__UpperCAmelCase , feature_extractor=__UpperCAmelCase ) __UpperCamelCase = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) __UpperCamelCase = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'facebook/detr-resnet-50' __UpperCamelCase = pipeline('object-detection' , model=__UpperCAmelCase ) __UpperCamelCase = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) __UpperCamelCase = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 0.9_9_8_5 __UpperCamelCase = 'facebook/detr-resnet-50' __UpperCamelCase = pipeline('object-detection' , model=__UpperCAmelCase ) __UpperCamelCase = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=__UpperCAmelCase ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) @require_torch @require_pytesseract @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'Narsil/layoutlmv3-finetuned-funsd' __UpperCamelCase = 0.9_9_9_3 __UpperCamelCase = pipeline('object-detection' , model=__UpperCAmelCase , threshold=__UpperCAmelCase ) __UpperCamelCase = object_detector( 'https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png' ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, ] , )
316
"""simple docstring""" def A ( snake_case :int ) -> int: __UpperCamelCase = [1] __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 0, 0, 0 __UpperCamelCase = ugly_nums[ia] * 2 __UpperCamelCase = ugly_nums[ia] * 3 __UpperCamelCase = ugly_nums[ia] * 5 for _ in range(1 , snake_case ): __UpperCamelCase = min(snake_case , snake_case , snake_case ) ugly_nums.append(snake_case ) if next_num == next_a: ia += 1 __UpperCamelCase = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 __UpperCamelCase = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 __UpperCamelCase = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(2_0_0) = }''')
316
1
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path UpperCamelCase_ = [ {"dataset": "wikipedia", "config_name": "20220301.de"}, {"dataset": "wikipedia", "config_name": "20220301.en"}, {"dataset": "wikipedia", "config_name": "20220301.fr"}, {"dataset": "wikipedia", "config_name": "20220301.frr"}, {"dataset": "wikipedia", "config_name": "20220301.it"}, {"dataset": "wikipedia", "config_name": "20220301.simple"}, {"dataset": "snli", "config_name": "plain_text"}, {"dataset": "eli5", "config_name": "LFQA_reddit"}, {"dataset": "wiki40b", "config_name": "en"}, {"dataset": "wiki_dpr", "config_name": "psgs_w100.nq.compressed"}, {"dataset": "wiki_dpr", "config_name": "psgs_w100.nq.no_index"}, {"dataset": "wiki_dpr", "config_name": "psgs_w100.multiset.no_index"}, {"dataset": "natural_questions", "config_name": "default"}, ] def lowercase__( __UpperCamelCase: Optional[int]=True ): """simple docstring""" if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=SCREAMING_SNAKE_CASE ) ) class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' A : Optional[Any] = None A : Optional[Any] = None def UpperCamelCase_ ( self, A, A ): '''simple docstring''' with TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : Union[str, Any] = dataset_module_factory(A, cache_dir=A ) SCREAMING_SNAKE_CASE : Optional[Any] = import_main_class(dataset_module.module_path, dataset=A ) SCREAMING_SNAKE_CASE : DatasetBuilder = builder_cls( cache_dir=A, config_name=A, hash=dataset_module.hash, ) SCREAMING_SNAKE_CASE : Union[str, Any] = '/'.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A ).replace(os.sep, '/' ), config.DATASET_INFO_FILENAME, ] ) SCREAMING_SNAKE_CASE : int = cached_path(A, cache_dir=A ) self.assertTrue(os.path.exists(A ) ) @pytest.mark.integration def lowercase__( __UpperCamelCase: Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = tmp_path_factory.mktemp('test_hf_gcp' ) / 'test_wikipedia_simple' SCREAMING_SNAKE_CASE : Dict = dataset_module_factory('wikipedia' ,cache_dir=__UpperCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = import_main_class(dataset_module.module_path ) SCREAMING_SNAKE_CASE : DatasetBuilder = builder_cls( cache_dir=__UpperCamelCase ,config_name='20220301.frr' ,hash=dataset_module.hash ,) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam SCREAMING_SNAKE_CASE : int = None builder_instance.download_and_prepare() SCREAMING_SNAKE_CASE : int = builder_instance.as_dataset() assert ds @pytest.mark.integration def lowercase__( __UpperCamelCase: List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = dataset_module_factory('wikipedia' ,cache_dir=__UpperCamelCase ) SCREAMING_SNAKE_CASE : int = import_main_class(dataset_module.module_path ,dataset=__UpperCamelCase ) SCREAMING_SNAKE_CASE : DatasetBuilder = builder_cls( cache_dir=__UpperCamelCase ,config_name='20220301.frr' ,hash=dataset_module.hash ,) SCREAMING_SNAKE_CASE : List[str] = builder_instance.as_streaming_dataset() assert ds assert isinstance(__UpperCamelCase ,__UpperCamelCase ) assert "train" in ds assert isinstance(ds['train'] ,__UpperCamelCase ) assert next(iter(ds['train'] ) )
362
'''simple docstring''' from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging UpperCamelCase_ = logging.get_logger(__name__) class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' A : int = ['''pixel_values'''] def __init__( self, A = True, A = None, A = PILImageResampling.BICUBIC, A = True, A = None, A = True, A = 1 / 255, A = True, A = IMAGENET_DEFAULT_MEAN, A = IMAGENET_DEFAULT_STD, **A, ): '''simple docstring''' super().__init__(**A ) SCREAMING_SNAKE_CASE : Any = size if size is not None else {'shortest_edge': 224} SCREAMING_SNAKE_CASE : List[str] = get_size_dict(A, default_to_square=A ) SCREAMING_SNAKE_CASE : str = crop_size if crop_size is not None else {'height': 224, 'width': 224} SCREAMING_SNAKE_CASE : int = get_size_dict(A, param_name='crop_size' ) SCREAMING_SNAKE_CASE : Any = do_resize SCREAMING_SNAKE_CASE : Dict = size SCREAMING_SNAKE_CASE : int = resample SCREAMING_SNAKE_CASE : List[Any] = do_center_crop SCREAMING_SNAKE_CASE : Dict = crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = do_rescale SCREAMING_SNAKE_CASE : Any = rescale_factor SCREAMING_SNAKE_CASE : List[Any] = do_normalize SCREAMING_SNAKE_CASE : List[str] = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN SCREAMING_SNAKE_CASE : Optional[int] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCamelCase_ ( self, A, A, A = PILImageResampling.BICUBIC, A = None, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = get_size_dict(A, default_to_square=A ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: SCREAMING_SNAKE_CASE : List[str] = int((256 / 224) * size['shortest_edge'] ) SCREAMING_SNAKE_CASE : int = get_resize_output_image_size(A, size=A, default_to_square=A ) SCREAMING_SNAKE_CASE : List[str] = {'height': output_size[0], 'width': output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( A, size=(size_dict['height'], size_dict['width']), resample=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A = None, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(A, size=(size['height'], size['width']), data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A = None, **A, ): '''simple docstring''' return rescale(A, scale=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A, A = None, **A, ): '''simple docstring''' return normalize(A, mean=A, std=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = ChannelDimension.FIRST, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Optional[int] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : List[Any] = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : Any = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : str = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Dict = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : int = size if size is not None else self.size SCREAMING_SNAKE_CASE : Dict = get_size_dict(A, default_to_square=A ) SCREAMING_SNAKE_CASE : Union[str, Any] = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Optional[Any] = get_size_dict(A, param_name='crop_size' ) SCREAMING_SNAKE_CASE : List[Any] = make_list_of_images(A ) if not valid_images(A ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : Optional[int] = [to_numpy_array(A ) for image in images] if do_resize: SCREAMING_SNAKE_CASE : List[str] = [self.resize(A, A, A ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE : List[Any] = [self.center_crop(A, A ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE : List[str] = [self.rescale(A, A ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE : Optional[int] = [self.normalize(A, A, A ) for image in images] SCREAMING_SNAKE_CASE : List[str] = [to_channel_dimension_format(A, A ) for image in images] SCREAMING_SNAKE_CASE : Optional[int] = {'pixel_values': images} return BatchFeature(data=A, tensor_type=A )
246
0
'''simple docstring''' from __future__ import annotations import math def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> list: '''simple docstring''' if len(__UpperCAmelCase ) != 2 or len(a[0] ) != 2 or len(__UpperCAmelCase ) != 2 or len(b[0] ) != 2: raise Exception('''Matrices are not 2x2''' ) snake_case_ = [ [a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]], [a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]], ] return new_matrix def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Any: '''simple docstring''' return [ [matrix_a[row][col] + matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(__UpperCAmelCase ) ) ] def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> int: '''simple docstring''' return [ [matrix_a[row][col] - matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(__UpperCAmelCase ) ) ] def __magic_name__ ( __UpperCAmelCase ) -> tuple[list, list, list, list]: '''simple docstring''' if len(__UpperCAmelCase ) % 2 != 0 or len(a[0] ) % 2 != 0: raise Exception('''Odd matrices are not supported!''' ) snake_case_ = len(__UpperCAmelCase ) snake_case_ = matrix_length // 2 snake_case_ = [[a[i][j] for j in range(__UpperCAmelCase, __UpperCAmelCase )] for i in range(__UpperCAmelCase )] snake_case_ = [ [a[i][j] for j in range(__UpperCAmelCase, __UpperCAmelCase )] for i in range(__UpperCAmelCase, __UpperCAmelCase ) ] snake_case_ = [[a[i][j] for j in range(__UpperCAmelCase )] for i in range(__UpperCAmelCase )] snake_case_ = [[a[i][j] for j in range(__UpperCAmelCase )] for i in range(__UpperCAmelCase, __UpperCAmelCase )] return top_left, top_right, bot_left, bot_right def __magic_name__ ( __UpperCAmelCase ) -> tuple[int, int]: '''simple docstring''' return len(__UpperCAmelCase ), len(matrix[0] ) def __magic_name__ ( __UpperCAmelCase ) -> None: '''simple docstring''' print('''\n'''.join(str(__UpperCAmelCase ) for line in matrix ) ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> list: '''simple docstring''' if matrix_dimensions(__UpperCAmelCase ) == (2, 2): return default_matrix_multiplication(__UpperCAmelCase, __UpperCAmelCase ) snake_case_ ,snake_case_ ,snake_case_ ,snake_case_ = split_matrix(__UpperCAmelCase ) snake_case_ ,snake_case_ ,snake_case_ ,snake_case_ = split_matrix(__UpperCAmelCase ) snake_case_ = actual_strassen(__UpperCAmelCase, matrix_subtraction(__UpperCAmelCase, __UpperCAmelCase ) ) snake_case_ = actual_strassen(matrix_addition(__UpperCAmelCase, __UpperCAmelCase ), __UpperCAmelCase ) snake_case_ = actual_strassen(matrix_addition(__UpperCAmelCase, __UpperCAmelCase ), __UpperCAmelCase ) snake_case_ = actual_strassen(__UpperCAmelCase, matrix_subtraction(__UpperCAmelCase, __UpperCAmelCase ) ) snake_case_ = actual_strassen(matrix_addition(__UpperCAmelCase, __UpperCAmelCase ), matrix_addition(__UpperCAmelCase, __UpperCAmelCase ) ) snake_case_ = actual_strassen(matrix_subtraction(__UpperCAmelCase, __UpperCAmelCase ), matrix_addition(__UpperCAmelCase, __UpperCAmelCase ) ) snake_case_ = actual_strassen(matrix_subtraction(__UpperCAmelCase, __UpperCAmelCase ), matrix_addition(__UpperCAmelCase, __UpperCAmelCase ) ) snake_case_ = matrix_addition(matrix_subtraction(matrix_addition(__UpperCAmelCase, __UpperCAmelCase ), __UpperCAmelCase ), __UpperCAmelCase ) snake_case_ = matrix_addition(__UpperCAmelCase, __UpperCAmelCase ) snake_case_ = matrix_addition(__UpperCAmelCase, __UpperCAmelCase ) snake_case_ = matrix_subtraction(matrix_subtraction(matrix_addition(__UpperCAmelCase, __UpperCAmelCase ), __UpperCAmelCase ), __UpperCAmelCase ) # construct the new matrix from our 4 quadrants snake_case_ = [] for i in range(len(__UpperCAmelCase ) ): new_matrix.append(top_left[i] + top_right[i] ) for i in range(len(__UpperCAmelCase ) ): new_matrix.append(bot_left[i] + bot_right[i] ) return new_matrix def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> list: '''simple docstring''' if matrix_dimensions(__UpperCAmelCase )[1] != matrix_dimensions(__UpperCAmelCase )[0]: snake_case_ = ( '''Unable to multiply these matrices, please check the dimensions.\n''' F"Matrix A: {matrixa}\n" F"Matrix B: {matrixa}" ) raise Exception(__UpperCAmelCase ) snake_case_ = matrix_dimensions(__UpperCAmelCase ) snake_case_ = matrix_dimensions(__UpperCAmelCase ) if dimensiona[0] == dimensiona[1] and dimensiona[0] == dimensiona[1]: return [matrixa, matrixa] snake_case_ = max(*__UpperCAmelCase, *__UpperCAmelCase ) snake_case_ = int(math.pow(2, math.ceil(math.loga(__UpperCAmelCase ) ) ) ) snake_case_ = matrixa snake_case_ = matrixa # Adding zeros to the matrices so that the arrays dimensions are the same and also # power of 2 for i in range(0, __UpperCAmelCase ): if i < dimensiona[0]: for _ in range(dimensiona[1], __UpperCAmelCase ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) if i < dimensiona[0]: for _ in range(dimensiona[1], __UpperCAmelCase ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) snake_case_ = actual_strassen(__UpperCAmelCase, __UpperCAmelCase ) # Removing the additional zeros for i in range(0, __UpperCAmelCase ): if i < dimensiona[0]: for _ in range(dimensiona[1], __UpperCAmelCase ): final_matrix[i].pop() else: final_matrix.pop() return final_matrix if __name__ == "__main__": a : List[Any] = [ [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 2, 3, 1], ] a : Union[str, Any] = [[0, 2, 1, 1], [16, 2, 3, 3], [2, 2, 7, 7], [13, 11, 22, 4]] print(strassen(matrixa, matrixa))
56
def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = 1 for i in range(1 , num + 1 ): fact *= i return fact def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = 0 while number > 0: _lowerCAmelCase = number % 10 sum_of_digits += last_digit _lowerCAmelCase = number // 10 # Removing the last_digit from the given number return sum_of_digits def _UpperCAmelCase ( snake_case = 1_00 ): """simple docstring""" _lowerCAmelCase = factorial(snake_case ) _lowerCAmelCase = split_and_add(snake_case ) return result if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
82
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin _UpperCAmelCase : Union[str, Any] = False @skip_mps class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = StableDiffusionAttendAndExcitePipeline UpperCAmelCase__ = False UpperCAmelCase__ = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS.union({"""token_indices"""} ) UpperCAmelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def A_ ( cls : Optional[Any] ) -> Union[str, Any]: super().setUpClass() torch.use_deterministic_algorithms(UpperCAmelCase ) @classmethod def A_ ( cls : str ) -> Any: super().tearDownClass() torch.use_deterministic_algorithms(UpperCAmelCase ) def A_ ( self : str ) -> Dict: torch.manual_seed(0 ) lowerCamelCase__ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=UpperCAmelCase , ) lowerCamelCase__ : Dict = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' , clip_sample=UpperCAmelCase , set_alpha_to_one=UpperCAmelCase , ) torch.manual_seed(0 ) lowerCamelCase__ : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) lowerCamelCase__ : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) lowerCamelCase__ : Tuple = CLIPTextModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCamelCase__ : Optional[int] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self : str , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any]=0 ) -> Optional[int]: if str(UpperCAmelCase ).startswith('mps' ): lowerCamelCase__ : str = torch.manual_seed(UpperCAmelCase ) else: lowerCamelCase__ : Tuple = torch.Generator(device=UpperCAmelCase ).manual_seed(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = { 'prompt': 'a cat and a frog', 'token_indices': [2, 5], 'generator': generator, 'num_inference_steps': 1, 'guidance_scale': 6.0, 'output_type': 'numpy', 'max_iter_to_alter': 2, 'thresholds': {0: 0.7}, } return inputs def A_ ( self : List[str] ) -> Dict: lowerCamelCase__ : Optional[Any] = 'cpu' lowerCamelCase__ : Dict = self.get_dummy_components() lowerCamelCase__ : List[str] = self.pipeline_class(**UpperCAmelCase ) pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowerCamelCase__ : Any = self.get_dummy_inputs(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = pipe(**UpperCAmelCase ).images lowerCamelCase__ : Tuple = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) lowerCamelCase__ : List[str] = np.array( [0.6_3_9_0_5_3_6_4, 0.6_2_8_9_7_3_0_7, 0.4_8_5_9_9_0_1_7, 0.5_1_3_3_6_2_4, 0.5_5_5_0_0_4_8, 0.4_5_7_6_9_5_1_6, 0.5_0_3_2_6_9_7_3, 0.5_0_2_3_1_3_9, 0.4_5_3_8_4_4_9_6] ) lowerCamelCase__ : List[str] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(UpperCAmelCase , 1e-3 ) def A_ ( self : Optional[int] ) -> List[Any]: super().test_cpu_offload_forward_pass(expected_max_diff=5e-4 ) def A_ ( self : int ) -> int: # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def A_ ( self : Optional[Any] ) -> Dict: self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7e-4 ) def A_ ( self : Dict ) -> int: super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def A_ ( self : List[str] ) -> Tuple: super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5e-4 ) def A_ ( self : List[Any] ) -> Optional[int]: super().test_save_load_local(expected_max_difference=5e-4 ) def A_ ( self : List[Any] ) -> Union[str, Any]: super().test_save_load_optional_components(expected_max_difference=4e-4 ) @require_torch_gpu @slow class lowerCAmelCase ( unittest.TestCase ): @classmethod def A_ ( cls : str ) -> Tuple: super().setUpClass() torch.use_deterministic_algorithms(UpperCAmelCase ) @classmethod def A_ ( cls : Any ) -> Union[str, Any]: super().tearDownClass() torch.use_deterministic_algorithms(UpperCAmelCase ) def A_ ( self : List[str] ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self : Optional[Any] ) -> Union[str, Any]: lowerCamelCase__ : Union[str, Any] = torch.manual_seed(51 ) lowerCamelCase__ : Any = StableDiffusionAttendAndExcitePipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , safety_checker=UpperCAmelCase , torch_dtype=torch.floataa ) pipe.to('cuda' ) lowerCamelCase__ : List[Any] = 'a painting of an elephant with glasses' lowerCamelCase__ : List[Any] = [5, 7] lowerCamelCase__ : Union[str, Any] = pipe( prompt=UpperCAmelCase , token_indices=UpperCAmelCase , guidance_scale=7.5 , generator=UpperCAmelCase , num_inference_steps=5 , max_iter_to_alter=5 , output_type='numpy' , ).images[0] lowerCamelCase__ : List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy' ) assert np.abs((expected_image - image).max() ) < 5e-1
45
from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Union[str, Any] = { """google/efficientnet-b7""": """https://huggingface.co/google/efficientnet-b7/resolve/main/config.json""", } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """efficientnet""" def __init__( self : Tuple , UpperCAmelCase : int = 3 , UpperCAmelCase : int = 600 , UpperCAmelCase : float = 2.0 , UpperCAmelCase : float = 3.1 , UpperCAmelCase : int = 8 , UpperCAmelCase : List[int] = [3, 3, 5, 3, 5, 5, 3] , UpperCAmelCase : List[int] = [32, 16, 24, 40, 80, 112, 192] , UpperCAmelCase : List[int] = [16, 24, 40, 80, 112, 192, 320] , UpperCAmelCase : List[int] = [] , UpperCAmelCase : List[int] = [1, 2, 2, 2, 1, 2, 1] , UpperCAmelCase : List[int] = [1, 2, 2, 3, 3, 4, 1] , UpperCAmelCase : List[int] = [1, 6, 6, 6, 6, 6, 6] , UpperCAmelCase : float = 0.2_5 , UpperCAmelCase : str = "swish" , UpperCAmelCase : int = 2560 , UpperCAmelCase : str = "mean" , UpperCAmelCase : float = 0.0_2 , UpperCAmelCase : float = 0.0_0_1 , UpperCAmelCase : float = 0.9_9 , UpperCAmelCase : float = 0.5 , UpperCAmelCase : float = 0.2 , **UpperCAmelCase : int , ) -> Any: super().__init__(**UpperCAmelCase ) lowerCamelCase__ : List[Any] = num_channels lowerCamelCase__ : List[str] = image_size lowerCamelCase__ : Union[str, Any] = width_coefficient lowerCamelCase__ : Optional[Any] = depth_coefficient lowerCamelCase__ : Union[str, Any] = depth_divisor lowerCamelCase__ : Dict = kernel_sizes lowerCamelCase__ : Union[str, Any] = in_channels lowerCamelCase__ : Dict = out_channels lowerCamelCase__ : Dict = depthwise_padding lowerCamelCase__ : int = strides lowerCamelCase__ : List[str] = num_block_repeats lowerCamelCase__ : Optional[Any] = expand_ratios lowerCamelCase__ : List[str] = squeeze_expansion_ratio lowerCamelCase__ : int = hidden_act lowerCamelCase__ : int = hidden_dim lowerCamelCase__ : int = pooling_type lowerCamelCase__ : Optional[Any] = initializer_range lowerCamelCase__ : Any = batch_norm_eps lowerCamelCase__ : List[Any] = batch_norm_momentum lowerCamelCase__ : int = dropout_rate lowerCamelCase__ : int = drop_connect_rate lowerCamelCase__ : List[Any] = sum(UpperCAmelCase ) * 4 class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = version.parse("""1.11""" ) @property def A_ ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def A_ ( self : List[Any] ) -> float: return 1e-5
45
1
"""simple docstring""" import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __A = "▁" __A = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" _UpperCAmelCase :Union[str, Any] = BigBirdTokenizer _UpperCAmelCase :Tuple = BigBirdTokenizerFast _UpperCAmelCase :str = True _UpperCAmelCase :Tuple = True def _snake_case ( self ): super().setUp() lowercase__: Optional[Any] = self.tokenizer_class(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self ): lowercase__: List[str] = '''<s>''' lowercase__: Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''[MASK]''' ) self.assertEqual(len(_UpperCAmelCase ) , 1004 ) def _snake_case ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def _snake_case ( self ): if not self.test_rust_tokenizer: return lowercase__: Tuple = self.get_tokenizer() lowercase__: Tuple = self.get_rust_tokenizer() lowercase__: Union[str, Any] = '''I was born in 92000, and this is falsé.''' lowercase__: int = tokenizer.tokenize(_UpperCAmelCase ) lowercase__: Optional[Any] = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Tuple = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) lowercase__: List[str] = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Any = self.get_rust_tokenizer() lowercase__: str = tokenizer.encode(_UpperCAmelCase ) lowercase__: Union[str, Any] = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: Dict = BigBirdTokenizer(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) lowercase__: Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowercase__: Dict = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowercase__: List[Any] = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) self.assertListEqual( _UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowercase__: Any = tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual( _UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def _snake_case ( self ): return BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) @slow def _snake_case ( self ): lowercase__: Dict = '''Hello World!''' lowercase__: List[Any] = [65, 18536, 2260, 101, 66] self.assertListEqual(_UpperCAmelCase , self.big_tokenizer.encode(_UpperCAmelCase ) ) @slow def _snake_case ( self ): lowercase__: str = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) # fmt: off lowercase__: Optional[Any] = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(_UpperCAmelCase , self.big_tokenizer.encode(_UpperCAmelCase ) ) @require_torch @slow def _snake_case ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence lowercase__: Any = list(self.big_tokenizer.get_vocab().keys() )[:10] lowercase__: Dict = ''' '''.join(_UpperCAmelCase ) lowercase__: Union[str, Any] = self.big_tokenizer.encode_plus(_UpperCAmelCase , return_tensors='''pt''' , return_token_type_ids=_UpperCAmelCase ) lowercase__: str = self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=_UpperCAmelCase ) lowercase__: List[Any] = BigBirdConfig(attention_type='''original_full''' ) lowercase__: Any = BigBirdModel(_UpperCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_UpperCAmelCase ) model(**_UpperCAmelCase ) @slow def _snake_case ( self ): lowercase__: Dict = BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) lowercase__: Dict = tokenizer.decode(tokenizer('''Paris is the [MASK].''' ).input_ids ) self.assertTrue(decoded_text == '''[CLS] Paris is the[MASK].[SEP]''' ) @slow def _snake_case ( self ): # fmt: off lowercase__: Any = {'''input_ids''': [[65, 39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114, 66], [65, 448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='''google/bigbird-roberta-base''' , revision='''215c99f1600e06f83acce68422f2035b2b5c3510''' , )
177
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig __A = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring __A = "UperNetConfig" class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 0 , _UpperCAmelCase = False , _UpperCAmelCase = 1 , ): super().__init__() lowercase__: List[Any] = nn.Convad( in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , bias=_UpperCAmelCase , dilation=_UpperCAmelCase , ) lowercase__: List[Any] = nn.BatchNormad(_UpperCAmelCase ) lowercase__: int = nn.ReLU() def _snake_case ( self , _UpperCAmelCase ): lowercase__: Dict = self.conv(_UpperCAmelCase ) lowercase__: Optional[int] = self.batch_norm(_UpperCAmelCase ) lowercase__: List[Any] = self.activation(_UpperCAmelCase ) return output class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: int = [ nn.AdaptiveAvgPoolad(_UpperCAmelCase ), UperNetConvModule(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase ): lowercase__: Any = input for layer in self.layers: lowercase__: Any = layer(_UpperCAmelCase ) return hidden_state class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: int = pool_scales lowercase__: Optional[Any] = align_corners lowercase__: Optional[int] = in_channels lowercase__: Optional[Any] = channels lowercase__: List[Any] = [] for i, pool_scale in enumerate(_UpperCAmelCase ): lowercase__: Optional[int] = UperNetPyramidPoolingBlock(pool_scale=_UpperCAmelCase , in_channels=_UpperCAmelCase , channels=_UpperCAmelCase ) self.blocks.append(_UpperCAmelCase ) self.add_module(str(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase ): lowercase__: Union[str, Any] = [] for ppm in self.blocks: lowercase__: Tuple = ppm(_UpperCAmelCase ) lowercase__: Any = nn.functional.interpolate( _UpperCAmelCase , size=x.size()[2:] , mode='''bilinear''' , align_corners=self.align_corners ) ppm_outs.append(_UpperCAmelCase ) return ppm_outs class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: Optional[int] = config lowercase__: int = config.pool_scales # e.g. (1, 2, 3, 6) lowercase__: Optional[int] = in_channels lowercase__: List[str] = config.hidden_size lowercase__: List[str] = False lowercase__: List[str] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module lowercase__: Dict = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) lowercase__: int = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module lowercase__: List[Any] = nn.ModuleList() lowercase__: Union[str, Any] = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer lowercase__: int = UperNetConvModule(_UpperCAmelCase , self.channels , kernel_size=1 ) lowercase__: Dict = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(_UpperCAmelCase ) self.fpn_convs.append(_UpperCAmelCase ) lowercase__: Any = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def _snake_case ( self ): self.apply(self._init_weights ) def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def _snake_case ( self , _UpperCAmelCase ): lowercase__: List[str] = inputs[-1] lowercase__: str = [x] psp_outs.extend(self.psp_modules(_UpperCAmelCase ) ) lowercase__: Dict = torch.cat(_UpperCAmelCase , dim=1 ) lowercase__: Tuple = self.bottleneck(_UpperCAmelCase ) return output def _snake_case ( self , _UpperCAmelCase ): # build laterals lowercase__: Dict = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(_UpperCAmelCase ) ) # build top-down path lowercase__: int = len(_UpperCAmelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowercase__: str = laterals[i - 1].shape[2:] lowercase__: Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=_UpperCAmelCase , mode='''bilinear''' , align_corners=self.align_corners ) # build outputs lowercase__: str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowercase__: Any = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='''bilinear''' , align_corners=self.align_corners ) lowercase__: int = torch.cat(_UpperCAmelCase , dim=1 ) lowercase__: Tuple = self.fpn_bottleneck(_UpperCAmelCase ) lowercase__: Dict = self.classifier(_UpperCAmelCase ) return output class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase = 2 , _UpperCAmelCase = 3 , _UpperCAmelCase = 1 ): super().__init__() lowercase__: Optional[Any] = config lowercase__: Optional[Any] = config.auxiliary_in_channels lowercase__: List[Any] = config.auxiliary_channels lowercase__: Tuple = config.auxiliary_num_convs lowercase__: Any = config.auxiliary_concat_input lowercase__: Optional[int] = in_index lowercase__: Tuple = (kernel_size // 2) * dilation lowercase__: Tuple = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , dilation=_UpperCAmelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , dilation=_UpperCAmelCase ) ) if self.num_convs == 0: lowercase__: List[Any] = nn.Identity() else: lowercase__: Union[str, Any] = nn.Sequential(*_UpperCAmelCase ) if self.concat_input: lowercase__: Dict = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=_UpperCAmelCase , padding=kernel_size // 2 ) lowercase__: Union[str, Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def _snake_case ( self ): self.apply(self._init_weights ) def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def _snake_case ( self , _UpperCAmelCase ): # just take the relevant feature maps lowercase__: Dict = encoder_hidden_states[self.in_index] lowercase__: Optional[int] = self.convs(_UpperCAmelCase ) if self.concat_input: lowercase__: Optional[int] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) lowercase__: Dict = self.classifier(_UpperCAmelCase ) return output class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :Dict = UperNetConfig _UpperCAmelCase :int = "pixel_values" _UpperCAmelCase :Optional[Any] = True def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def _snake_case ( self ): self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase=False ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowercase__: Any = value __A = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." ,_UpperCAmelCase ,) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase ): super().__init__(_UpperCAmelCase ) lowercase__: Optional[int] = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) lowercase__: Any = UperNetHead(_UpperCAmelCase , in_channels=self.backbone.channels ) lowercase__: Tuple = UperNetFCNHead(_UpperCAmelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('''batch_size, sequence_length''' ) ) @replace_return_docstrings(output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC ) def _snake_case ( self , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , ): lowercase__: Tuple = return_dict if return_dict is not None else self.config.use_return_dict lowercase__: Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__: str = output_attentions if output_attentions is not None else self.config.output_attentions lowercase__: List[str] = self.backbone.forward_with_filtered_kwargs( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , output_attentions=_UpperCAmelCase ) lowercase__: Tuple = outputs.feature_maps lowercase__: Union[str, Any] = self.decode_head(_UpperCAmelCase ) lowercase__: str = nn.functional.interpolate(_UpperCAmelCase , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=_UpperCAmelCase ) lowercase__: Any = None if self.auxiliary_head is not None: lowercase__: Union[str, Any] = self.auxiliary_head(_UpperCAmelCase ) lowercase__: Tuple = nn.functional.interpolate( _UpperCAmelCase , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=_UpperCAmelCase ) lowercase__: List[Any] = None if labels is not None: if self.config.num_labels == 1: raise ValueError('''The number of labels should be greater than one''' ) else: # compute weighted loss lowercase__: List[str] = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) lowercase__: Optional[Any] = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Dict = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: int = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: lowercase__: Tuple = (logits,) + outputs[1:] else: lowercase__: Optional[int] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=_UpperCAmelCase , logits=_UpperCAmelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
177
1
def A(__a: int , __a: int ): return int((input_a, input_a).count(0 ) == 0 ) def A(): assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
365
import math def A(__a: int ): return math.sqrt(__a ) * math.sqrt(__a ) == num def A(__a: int ): lowerCAmelCase_ = 0 lowerCAmelCase_ = n while left <= right: lowerCAmelCase_ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: lowerCAmelCase_ = mid - 1 else: lowerCAmelCase_ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
22
0
import math def __UpperCamelCase ( _A : int , _A : List[str] ) ->Tuple: """simple docstring""" if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(_A ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("""This should never happen""" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. __A : List[Any] = 'Enter the base and the power separated by a comma: ' __A, __A : List[Any] = map(int, input(prompt).split(',')) __A, __A : str = map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. __A : str = res(xa, ya) __A : Any = res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
154
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Any = logging.get_logger(__name__) __A : Dict = {'vocab_file': 'spiece.model'} __A : List[Any] = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<sep>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE="<cls>" , _SCREAMING_SNAKE_CASE="<mask>" , _SCREAMING_SNAKE_CASE=["<eop>", "<eod>"] , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , )-> None: lowerCamelCase_ =AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else mask_token lowerCamelCase_ ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_SCREAMING_SNAKE_CASE , remove_space=_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **_SCREAMING_SNAKE_CASE , ) lowerCamelCase_ =3 lowerCamelCase_ =do_lower_case lowerCamelCase_ =remove_space lowerCamelCase_ =keep_accents lowerCamelCase_ =vocab_file lowerCamelCase_ =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_SCREAMING_SNAKE_CASE ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) lowerCamelCase_ =jieba lowerCamelCase_ =str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _snake_case ( self )-> Any: return len(self.sp_model ) def _snake_case ( self )-> Dict: lowerCamelCase_ ={self.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self )-> List[Any]: lowerCamelCase_ =self.__dict__.copy() lowerCamelCase_ =None return state def __setstate__( self , _SCREAMING_SNAKE_CASE )-> List[Any]: lowerCamelCase_ =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase_ ={} lowerCamelCase_ =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> int: if self.remove_space: lowerCamelCase_ =""" """.join(inputs.strip().split() ) else: lowerCamelCase_ =inputs lowerCamelCase_ =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: lowerCamelCase_ =unicodedata.normalize("""NFKD""" , _SCREAMING_SNAKE_CASE ) lowerCamelCase_ ="""""".join([c for c in outputs if not unicodedata.combining(_SCREAMING_SNAKE_CASE )] ) if self.do_lower_case: lowerCamelCase_ =outputs.lower() return outputs def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> List[str]: lowerCamelCase_ =self.preprocess_text(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =self.sp_model.encode(_SCREAMING_SNAKE_CASE , out_type=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =[] for piece in pieces: if len(_SCREAMING_SNAKE_CASE ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowerCamelCase_ =self.sp_model.EncodeAsPieces(piece[:-1].replace(_SCREAMING_SNAKE_CASE , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase_ =cur_pieces[1:] else: lowerCamelCase_ =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_SCREAMING_SNAKE_CASE ) else: new_pieces.append(_SCREAMING_SNAKE_CASE ) return new_pieces def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: return self.sp_model.PieceToId(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: return self.sp_model.IdToPiece(_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: lowerCamelCase_ ="""""".join(_SCREAMING_SNAKE_CASE ).replace(_SCREAMING_SNAKE_CASE , """ """ ).strip() return out_string def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> List[int]: lowerCamelCase_ =[self.sep_token_id] lowerCamelCase_ =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False )-> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_SCREAMING_SNAKE_CASE , token_ids_a=_SCREAMING_SNAKE_CASE , already_has_special_tokens=_SCREAMING_SNAKE_CASE ) if token_ids_a is not None: return ([0] * len(_SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(_SCREAMING_SNAKE_CASE )) + [1, 1] return ([0] * len(_SCREAMING_SNAKE_CASE )) + [1, 1] def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> List[int]: lowerCamelCase_ =[self.sep_token_id] lowerCamelCase_ =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None )-> Tuple[str]: if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowerCamelCase_ =os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(_SCREAMING_SNAKE_CASE , """wb""" ) as fi: lowerCamelCase_ =self.sp_model.serialized_model_proto() fi.write(_SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _snake_case ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )-> Any: lowerCamelCase_ =super()._decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
154
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCAmelCase_ ( self ) -> Optional[Any]: torch.manual_seed(0 ) A_ : List[Any] = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return model @property def UpperCAmelCase_ ( self ) -> Union[str, Any]: torch.manual_seed(0 ) A_ : List[Any] = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , cross_attention_dim=10 , ) return model @property def UpperCAmelCase_ ( self ) -> Union[str, Any]: torch.manual_seed(0 ) A_ : str = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""DownEncoderBlock2D""", """DownEncoderBlock2D""") , up_block_types=("""UpDecoderBlock2D""", """UpDecoderBlock2D""") , ) A_ : Dict = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return vqvae, unet @slow def UpperCAmelCase_ ( self ) -> int: A_ : Tuple = """cpu""" # ensure determinism for the device-dependent torch.Generator A_ : List[Any] = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) A_ : int = DDPMScheduler() A_ : str = AudioDiffusionPipeline(vqvae=_lowerCamelCase , unet=self.dummy_unet , mel=_lowerCamelCase , scheduler=_lowerCamelCase ) A_ : List[Any] = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) A_ : Union[str, Any] = torch.Generator(device=_lowerCamelCase ).manual_seed(42 ) A_ : int = pipe(generator=_lowerCamelCase , steps=4 ) A_ : Optional[int] = output.audios[0] A_ : Union[str, Any] = output.images[0] A_ : Optional[int] = torch.Generator(device=_lowerCamelCase ).manual_seed(42 ) A_ : Tuple = pipe(generator=_lowerCamelCase , steps=4 , return_dict=_lowerCamelCase ) A_ : Tuple = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) A_ : Any = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] A_ : Optional[int] = np.frombuffer(image_from_tuple.tobytes() , dtype="""uint8""" )[:10] A_ : int = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 A_ : Tuple = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) A_ : Union[str, Any] = DDIMScheduler() A_ : Union[str, Any] = self.dummy_vqvae_and_unet A_ : Union[str, Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_lowerCamelCase , scheduler=_lowerCamelCase ) A_ : Any = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) np.random.seed(0 ) A_ : List[str] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) A_ : List[Any] = torch.Generator(device=_lowerCamelCase ).manual_seed(42 ) A_ : Tuple = pipe(raw_audio=_lowerCamelCase , generator=_lowerCamelCase , start_step=5 , steps=10 ) A_ : Optional[int] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) A_ : Optional[int] = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] A_ : Optional[Any] = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 A_ : Optional[int] = self.dummy_unet_condition A_ : List[str] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_lowerCamelCase , mel=_lowerCamelCase , scheduler=_lowerCamelCase ) A_ : Union[str, Any] = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) np.random.seed(0 ) A_ : Tuple = torch.rand((1, 1, 10) ) A_ : Optional[Any] = pipe(generator=_lowerCamelCase , encoding=_lowerCamelCase ) A_ : Optional[int] = output.images[0] A_ : Union[str, Any] = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] A_ : Dict = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ) -> List[str]: A_ : int = torch_device A_ : List[str] = DiffusionPipeline.from_pretrained("""teticio/audio-diffusion-ddim-256""" ) A_ : Tuple = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) A_ : Union[str, Any] = torch.Generator(device=_lowerCamelCase ).manual_seed(42 ) A_ : Any = pipe(generator=_lowerCamelCase ) A_ : List[Any] = output.audios[0] A_ : Optional[Any] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] A_ : Optional[Any] = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] A_ : int = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
164
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Any class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase ) -> Optional[Any]: A_ : Any = data A_ : Node | None = None class _lowerCAmelCase : """simple docstring""" def __init__( self ) -> List[str]: A_ : Tuple = None A_ : str = None def __iter__( self ) -> Iterator[Any]: A_ : Dict = self.head while self.head: yield node.data A_ : Optional[Any] = node.next if node == self.head: break def __len__( self ) -> int: return sum(1 for _ in self ) def __repr__( self ) -> str: return "->".join(str(_lowerCamelCase ) for item in iter(self ) ) def UpperCAmelCase_ ( self , _lowerCamelCase ) -> None: self.insert_nth(len(self ) , _lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase ) -> None: self.insert_nth(0 , _lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> None: if index < 0 or index > len(self ): raise IndexError("""list index out of range.""" ) A_ : Optional[int] = Node(_lowerCamelCase ) if self.head is None: A_ : str = new_node # first node points itself A_ : Union[str, Any] = new_node elif index == 0: # insert at head A_ : List[Any] = self.head A_ : List[Any] = new_node else: A_ : List[str] = self.head for _ in range(index - 1 ): A_ : Optional[int] = temp.next A_ : Tuple = temp.next A_ : str = new_node if index == len(self ) - 1: # insert at tail A_ : Optional[int] = new_node def UpperCAmelCase_ ( self ) -> List[Any]: return self.delete_nth(0 ) def UpperCAmelCase_ ( self ) -> Any: return self.delete_nth(len(self ) - 1 ) def UpperCAmelCase_ ( self , _lowerCamelCase = 0 ) -> Any: if not 0 <= index < len(self ): raise IndexError("""list index out of range.""" ) A_ : int = self.head if self.head == self.tail: # just one node A_ : int = None elif index == 0: # delete head node A_ : Union[str, Any] = self.tail.next.next A_ : Tuple = self.head.next else: A_ : Optional[int] = self.head for _ in range(index - 1 ): A_ : Tuple = temp.next A_ : Any = temp.next A_ : Tuple = temp.next.next if index == len(self ) - 1: # delete at tail A_ : List[str] = temp return delete_node.data def UpperCAmelCase_ ( self ) -> bool: return len(self ) == 0 def UpperCAmelCase ( ) -> None: """simple docstring""" A_ : Any = CircularLinkedList() assert len(a_ ) == 0 assert circular_linked_list.is_empty() is True assert str(a_ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(a_ ) == i circular_linked_list.insert_nth(a_ , i + 1 ) assert str(a_ ) == "->".join(str(a_ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(a_ ) == "->".join(str(a_ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(a_ ) == "->".join(str(a_ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(a_ ) == "->".join(str(a_ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(a_ ) == "->".join(str(a_ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
164
1
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase : Union[str, Any] = TypeVar("T") UpperCAmelCase : Dict = Union[List[T], Tuple[T, ...]] UpperCAmelCase : int = Union[T, List[T], Dict[str, T]] UpperCAmelCase : Tuple = Union[str, bytes, os.PathLike]
136
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase : Union[str, Any] = TypeVar("T") UpperCAmelCase : Dict = Union[List[T], Tuple[T, ...]] UpperCAmelCase : int = Union[T, List[T], Dict[str, T]] UpperCAmelCase : Tuple = Union[str, bytes, os.PathLike]
136
1
def lowerCamelCase__ ( a = 1_00_00_00 ) -> int: _A: Optional[Any] = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , a ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
301
import os from pathlib import Path def lowerCamelCase__ ( ) -> Optional[Any]: from torch.utils.cpp_extension import load _A: str = Path(a ).resolve().parent.parent.parent / '''kernels''' / '''deformable_detr''' _A: Tuple = [ root / filename for filename in [ '''vision.cpp''', os.path.join('''cpu''' , '''ms_deform_attn_cpu.cpp''' ), os.path.join('''cuda''' , '''ms_deform_attn_cuda.cu''' ), ] ] load( '''MultiScaleDeformableAttention''' , a , with_cuda=a , extra_include_paths=[str(a )] , extra_cflags=['''-DWITH_CUDA=1'''] , extra_cuda_cflags=[ '''-DCUDA_HAS_FP16=1''', '''-D__CUDA_NO_HALF_OPERATORS__''', '''-D__CUDA_NO_HALF_CONVERSIONS__''', '''-D__CUDA_NO_HALF2_OPERATORS__''', ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
301
1
'''simple docstring''' import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): lowercase = BioGptTokenizer lowercase = False def snake_case_ ( self ) -> Optional[int]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A_ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] A_ = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) A_ = ["l o 123", "lo w 1456", "e r</w> 1789", ""] A_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(UpperCAmelCase_ ) ) def snake_case_ ( self , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' A_ = "lower newer" A_ = "lower newer" return input_text, output_text def snake_case_ ( self ) -> Optional[int]: '''simple docstring''' A_ = BioGptTokenizer(self.vocab_file , self.merges_file ) A_ = "lower" A_ = ["low", "er</w>"] A_ = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) A_ = tokens + ["<unk>"] A_ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , UpperCAmelCase_ ) @slow def snake_case_ ( self ) -> Dict: '''simple docstring''' A_ = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A_ = tokenizer.encode("""sequence builders""" , add_special_tokens=UpperCAmelCase_ ) A_ = tokenizer.encode("""multi-sequence build""" , add_special_tokens=UpperCAmelCase_ ) A_ = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) A_ = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
162
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "ctc_proj", "mask_emb": "masked_spec_embed", } __A = [ "ctc_proj", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def lowerCAmelCase_ ( __a , __a , __a , __a , __a , __a ) -> Optional[Any]: """simple docstring""" for attribute in key.split("." ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models lowerCamelCase__: Optional[int] ="lm_head" lowerCamelCase__: Dict =getattr(__a , __a ) if weight_type is not None: lowerCamelCase__: str =getattr(__a , __a ).shape else: lowerCamelCase__: int =hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowerCamelCase__: Dict =value elif weight_type == "weight_g": lowerCamelCase__: Optional[Any] =value elif weight_type == "weight_v": lowerCamelCase__: int =value elif weight_type == "bias": lowerCamelCase__: List[str] =value else: lowerCamelCase__: Union[str, Any] =value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def lowerCAmelCase_ ( __a , __a , __a ) -> Any: """simple docstring""" lowerCamelCase__: List[Any] =[] lowerCamelCase__: List[str] =fairseq_model.state_dict() lowerCamelCase__: Optional[int] =hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): lowerCamelCase__: int =False if "conv_layers" in name: load_conv_layer( __a , __a , __a , __a , hf_model.config.feat_extract_norm == "group" , ) lowerCamelCase__: str =True else: for key, mapped_key in MAPPING.items(): lowerCamelCase__: List[str] ="unispeech." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: lowerCamelCase__: Optional[Any] =True if "*" in mapped_key: lowerCamelCase__: Optional[Any] =name.split(__a )[0].split("." )[-2] lowerCamelCase__: List[str] =mapped_key.replace("*" , __a ) if "weight_g" in name: lowerCamelCase__: List[str] ="weight_g" elif "weight_v" in name: lowerCamelCase__: Union[str, Any] ="weight_v" elif "bias" in name: lowerCamelCase__: Dict ="bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj lowerCamelCase__: Tuple ="weight" else: lowerCamelCase__: List[Any] =None set_recursively(__a , __a , __a , __a , __a , __a ) continue if not is_used: unused_weights.append(__a ) logger.warning(F"""Unused weights: {unused_weights}""" ) def lowerCAmelCase_ ( __a , __a , __a , __a , __a ) -> Union[str, Any]: """simple docstring""" lowerCamelCase__: Tuple =full_name.split("conv_layers." )[-1] lowerCamelCase__: List[str] =name.split("." ) lowerCamelCase__: str =int(items[0] ) lowerCamelCase__: Union[str, Any] =int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowerCamelCase__: List[str] =value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowerCamelCase__: Dict =value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) lowerCamelCase__: List[Any] =value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) lowerCamelCase__: List[str] =value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__a ) @torch.no_grad() def lowerCAmelCase_ ( __a , __a , __a=None , __a=None , __a=True ) -> int: """simple docstring""" if config_path is not None: lowerCamelCase__: str =UniSpeechConfig.from_pretrained(__a ) else: lowerCamelCase__: List[Any] =UniSpeechConfig() if is_finetuned: if dict_path: lowerCamelCase__: str =Dictionary.load_from_json(__a ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCamelCase__: Any =target_dict.pad_index lowerCamelCase__: int =target_dict.bos_index lowerCamelCase__: Any =target_dict.eos_index lowerCamelCase__: Dict =len(target_dict.symbols ) lowerCamelCase__: Optional[int] =os.path.join(__a , "vocab.json" ) if not os.path.isdir(__a ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(__a ) ) return os.makedirs(__a , exist_ok=__a ) lowerCamelCase__: Optional[Any] =target_dict.indices # fairseq has the <pad> and <s> switched lowerCamelCase__: Optional[Any] =42 lowerCamelCase__: List[Any] =43 with open(__a , "w" , encoding="utf-8" ) as vocab_handle: json.dump(__a , __a ) lowerCamelCase__: List[str] =WavaVecaPhonemeCTCTokenizer( __a , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=__a , ) lowerCamelCase__: Dict =True if config.feat_extract_norm == "layer" else False lowerCamelCase__: Tuple =WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=__a , return_attention_mask=__a , ) lowerCamelCase__: List[Any] =WavaVecaProcessor(feature_extractor=__a , tokenizer=__a ) processor.save_pretrained(__a ) lowerCamelCase__: int =UniSpeechForCTC(__a ) else: lowerCamelCase__: int =UniSpeechForPreTraining(__a ) if is_finetuned: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Optional[int] =fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] ), "w2v_path": checkpoint_path} ) else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Tuple =fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) lowerCamelCase__: List[str] =model[0].eval() recursively_load_weights(__a , __a , __a ) hf_unispeech.save_pretrained(__a ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) __A = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
10
0
"""simple docstring""" import re def SCREAMING_SNAKE_CASE__ ( snake_case : str )-> bool: '''simple docstring''' UpperCAmelCase__ : Dict = re.compile(r"^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$" ) if match := re.search(snake_case , snake_case ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator("""+918827897895"""))
298
"""simple docstring""" import numpy as np import datasets _lowerCAmelCase : Optional[int] = """ Compute the Mahalanobis Distance Mahalonobis distance is the distance between a point and a distribution. And not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance. It was introduced by Prof. P. C. Mahalanobis in 1936 and has been used in various statistical applications ever since [source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/] """ _lowerCAmelCase : Tuple = """\ @article{de2000mahalanobis, title={The mahalanobis distance}, author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L}, journal={Chemometrics and intelligent laboratory systems}, volume={50}, number={1}, pages={1--18}, year={2000}, publisher={Elsevier} } """ _lowerCAmelCase : Optional[int] = """ Args: X: List of datapoints to be compared with the `reference_distribution`. reference_distribution: List of datapoints from the reference distribution we want to compare to. Returns: mahalanobis: The Mahalonobis distance for each datapoint in `X`. Examples: >>> mahalanobis_metric = datasets.load_metric(\"mahalanobis\") >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]]) >>> print(results) {'mahalanobis': array([0.5])} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def __a ( self : Any ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence" ) , id="X" ), } ) , ) def __a ( self : Union[str, Any] , snake_case__ : List[str] , snake_case__ : Any ): '''simple docstring''' # convert to numpy arrays UpperCAmelCase__ : Union[str, Any] = np.array(snake_case__ ) UpperCAmelCase__ : Union[str, Any] = np.array(snake_case__ ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError("Expected `X` to be a 2D vector" ) if len(reference_distribution.shape ) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector" ) if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension" ) # Get mahalanobis distance for each prediction UpperCAmelCase__ : Optional[Any] = X - np.mean(snake_case__ ) UpperCAmelCase__ : Tuple = np.cov(reference_distribution.T ) try: UpperCAmelCase__ : str = np.linalg.inv(snake_case__ ) except np.linalg.LinAlgError: UpperCAmelCase__ : Optional[Any] = np.linalg.pinv(snake_case__ ) UpperCAmelCase__ : List[Any] = np.dot(snake_case__ , snake_case__ ) UpperCAmelCase__ : Tuple = np.dot(snake_case__ , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
298
1
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor snake_case : Union[str, Any] = logging.get_logger(__name__) class _snake_case ( snake_case ): def __init__( self , *_a , **_a ): warnings.warn( "The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ChineseCLIPImageProcessor instead." , _a , ) super().__init__(*_a , **_a )
281
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case : Optional[Any] = logging.get_logger(__name__) def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Union[str, Any]=False ) -> List[str]: '''simple docstring''' __magic_name__ : Union[str, Any] = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __magic_name__ : int = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Any , _snake_case : Dict=False ) -> int: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: __magic_name__ : int = "" else: __magic_name__ : Union[str, Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __magic_name__ : Optional[Any] = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) __magic_name__ : int = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ : Dict = in_proj_weight[ : config.hidden_size, : ] __magic_name__ : List[str] = in_proj_bias[: config.hidden_size] __magic_name__ : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __magic_name__ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __magic_name__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] __magic_name__ : int = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( _snake_case : List[str] ) -> List[str]: '''simple docstring''' __magic_name__ : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __magic_name__ : int = dct.pop(_snake_case ) __magic_name__ : List[Any] = val def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" __magic_name__ : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Any , _snake_case : int=False ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_snake_case , ) __magic_name__ : List[str] = ViTHybridConfig(backbone_config=_snake_case , image_size=384 , num_labels=1000 ) __magic_name__ : str = False # load original model from timm __magic_name__ : Union[str, Any] = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys __magic_name__ : List[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) __magic_name__ : Tuple = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) __magic_name__ : List[str] = "huggingface/label-files" __magic_name__ : int = "imagenet-1k-id2label.json" __magic_name__ : Optional[int] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="dataset" ) , "r" ) ) __magic_name__ : int = {int(_snake_case ): v for k, v in idalabel.items()} __magic_name__ : List[str] = idalabel __magic_name__ : List[str] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": __magic_name__ : List[str] = ViTHybridModel(_snake_case ).eval() else: __magic_name__ : str = ViTHybridForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # create image processor __magic_name__ : List[Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) __magic_name__ : int = transform.transforms __magic_name__ : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } __magic_name__ : int = ViTHybridImageProcessor( do_resize=_snake_case , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __magic_name__ : List[Any] = prepare_img() __magic_name__ : Any = transform(_snake_case ).unsqueeze(0 ) __magic_name__ : Tuple = processor(_snake_case , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): __magic_name__ : Optional[int] = model(_snake_case ) __magic_name__ : List[str] = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: __magic_name__ : List[str] = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1E-3 ) else: __magic_name__ : Any = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) snake_case : List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
1
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __SCREAMING_SNAKE_CASE ( unittest.TestCase): def UpperCamelCase__ ( self ): """simple docstring""" super().tearDown() gc.collect() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=_UpperCamelCase , dtype=jnp.bfloataa ) lowerCAmelCase__ , lowerCAmelCase__ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=_UpperCamelCase , from_pt=_UpperCamelCase , dtype=jnp.bfloataa ) lowerCAmelCase__ = controlnet_params lowerCAmelCase__ = 'bird' lowerCAmelCase__ = jax.device_count() lowerCAmelCase__ = pipe.prepare_text_inputs([prompts] * num_samples ) lowerCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) lowerCAmelCase__ = pipe.prepare_image_inputs([canny_image] * num_samples ) lowerCAmelCase__ = jax.random.PRNGKey(0 ) lowerCAmelCase__ = jax.random.split(_UpperCamelCase , jax.device_count() ) lowerCAmelCase__ = replicate(_UpperCamelCase ) lowerCAmelCase__ = shard(_UpperCamelCase ) lowerCAmelCase__ = shard(_UpperCamelCase ) lowerCAmelCase__ = pipe( prompt_ids=_UpperCamelCase , image=_UpperCamelCase , params=_UpperCamelCase , prng_seed=_UpperCamelCase , num_inference_steps=50 , jit=_UpperCamelCase , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowerCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase__ = images[0, 2_53:2_56, 2_53:2_56, -1] lowerCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase__ = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=_UpperCamelCase , dtype=jnp.bfloataa ) lowerCAmelCase__ , lowerCAmelCase__ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=_UpperCamelCase , from_pt=_UpperCamelCase , dtype=jnp.bfloataa ) lowerCAmelCase__ = controlnet_params lowerCAmelCase__ = 'Chef in the kitchen' lowerCAmelCase__ = jax.device_count() lowerCAmelCase__ = pipe.prepare_text_inputs([prompts] * num_samples ) lowerCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) lowerCAmelCase__ = pipe.prepare_image_inputs([pose_image] * num_samples ) lowerCAmelCase__ = jax.random.PRNGKey(0 ) lowerCAmelCase__ = jax.random.split(_UpperCamelCase , jax.device_count() ) lowerCAmelCase__ = replicate(_UpperCamelCase ) lowerCAmelCase__ = shard(_UpperCamelCase ) lowerCAmelCase__ = shard(_UpperCamelCase ) lowerCAmelCase__ = pipe( prompt_ids=_UpperCamelCase , image=_UpperCamelCase , params=_UpperCamelCase , prng_seed=_UpperCamelCase , num_inference_steps=50 , jit=_UpperCamelCase , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowerCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase__ = images[0, 2_53:2_56, 2_53:2_56, -1] lowerCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase__ = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
351
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup __snake_case : Any = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _UpperCamelCase ( UpperCamelCase_ : str = "mumbai" ) -> Generator[tuple[str, str], None, None]: """simple docstring""" lowerCAmelCase__ = BeautifulSoup(requests.get(url + location ).content , 'html.parser' ) # This attribute finds out all the specifics listed in a job for job in soup.find_all('div' , attrs={'data-tn-component': 'organicJob'} ): lowerCAmelCase__ = job.find('a' , attrs={'data-tn-element': 'jobTitle'} ).text.strip() lowerCAmelCase__ = job.find('span' , {'class': 'company'} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(f'Job {i:>2} is {job[0]} at {job[1]}')
122
0