code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" lowerCamelCase_ : int = """ # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git """ lowerCamelCase_ : Dict = [{"""type""": """code""", """content""": INSTALL_CONTENT}] lowerCamelCase_ : Union[str, Any] = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
81
def _SCREAMING_SNAKE_CASE ( a , a = 0 ) -> list: __A : int = length or len(a ) __A : str = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __A , __A : Optional[int] = list_data[i + 1], list_data[i] __A : Union[str, Any] = True return list_data if not swapped else bubble_sort(a , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A__ = logging.get_logger(__name__) A__ = { """EleutherAI/gpt-neox-20b""": """https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json""", # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class __lowerCAmelCase ( lowerCamelCase__ ): __lowerCamelCase = '''gpt_neox''' def __init__( self , _snake_case=50432 , _snake_case=6144 , _snake_case=44 , _snake_case=64 , _snake_case=24576 , _snake_case="gelu" , _snake_case=0.25 , _snake_case=10000 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.1 , _snake_case=2048 , _snake_case=0.02 , _snake_case=1e-5 , _snake_case=True , _snake_case=0 , _snake_case=2 , _snake_case=False , _snake_case=True , _snake_case=None , **_snake_case , ): """simple docstring""" super().__init__(bos_token_id=_snake_case , eos_token_id=_snake_case , **_snake_case ) _lowerCAmelCase = vocab_size _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = rotary_pct _lowerCAmelCase = rotary_emb_base _lowerCAmelCase = attention_dropout _lowerCAmelCase = hidden_dropout _lowerCAmelCase = classifier_dropout _lowerCAmelCase = initializer_range _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = use_cache _lowerCAmelCase = tie_word_embeddings _lowerCAmelCase = use_parallel_residual _lowerCAmelCase = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( """The hidden size is not divisble by the number of attention heads! Make sure to update them!""" ) def snake_case ( self ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _snake_case ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ F'got {self.rope_scaling}' ) _lowerCAmelCase = self.rope_scaling.get("""type""" , _snake_case ) _lowerCAmelCase = self.rope_scaling.get("""factor""" , _snake_case ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}' ) if rope_scaling_factor is None or not isinstance(_snake_case , _snake_case ) or rope_scaling_factor <= 1.0: raise ValueError(F'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}' )
82
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a ) -> int: if not nums: return 0 __A : Optional[int] = nums[0] __A : str = 0 for num in nums[1:]: __A , __A : Tuple = ( max_excluding + num, max(a , a ), ) return max(a , a ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' import os import pytest from attr import dataclass snake_case_ : Union[str, Any] = 'us-east-1' # defaults region @dataclass class lowercase__ : lowercase__ = 42 lowercase__ = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" lowercase__ = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 5_00, """save_steps""": 55_00, } lowercase__ = {**hyperparameters, """max_steps""": 10_00} @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCamelCase_ ( self : int ): '''simple docstring''' return F'{self.framework}-transfromers-test' @property def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return F'./tests/sagemaker/scripts/{self.framework}' @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Any = SageMakerTestEnvironment(framework=request.cls.framework )
83
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue_model_parallelism.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, ] ) class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Dict: if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="""utf-8""" , check=__A , ) assert hasattr(self , """env""" ) def __lowerCAmelCase ( self , __A ) -> Any: # configuration for running training on smdistributed Model Parallel lowerCAmelCase_ :Union[str, Any] = { """enabled""": True, """processes_per_host""": 8, } lowerCAmelCase_ :Tuple = { """enabled""": True, """parameters""": { """microbatches""": 4, """placement_strategy""": """spread""", """pipeline""": """interleaved""", """optimize""": """speed""", """partitions""": 4, """ddp""": True, }, } lowerCAmelCase_ :Any = {"""smdistributed""": {"""modelparallel""": smp_options}, """mpi""": mpi_options} lowerCAmelCase_ :Any = """trainer""" if self.script == """run_glue.py""" else """smtrainer""" # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-{instance_count}-smp-{name_extension}""" , instance_count=__A , instance_type=self.instance_type , debugger_hook_config=__A , hyperparameters={ **self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path, """max_steps""": 500, } , metric_definitions=self.env.metric_definitions , distribution=__A , py_version="""py36""" , ) def __lowerCAmelCase ( self , __A ) -> List[Any]: TrainingJobAnalytics(__A ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(1,)] ) def __lowerCAmelCase ( self , __A ) -> List[str]: # create estimator lowerCAmelCase_ :Any = self.create_estimator(__A ) # run training estimator.fit() # result dataframe lowerCAmelCase_ :Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis lowerCAmelCase_ :List[str] = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) lowerCAmelCase_ :Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping lowerCAmelCase_ :Optional[int] = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , __A )
84
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _SCREAMING_SNAKE_CASE : int = { "169M": 12, "430M": 24, "1B5": 24, "3B": 32, "7B": 32, "14B": 40, } _SCREAMING_SNAKE_CASE : Dict = { "169M": 768, "430M": 1024, "1B5": 2048, "3B": 2560, "7B": 4096, "14B": 5120, } def UpperCamelCase_( snake_case : Union[str, Any] ): '''simple docstring''' snake_case_ = list(state_dict.keys() ) for name in state_dict_keys: snake_case_ = state_dict.pop(snake_case ) # emb -> embedding if name.startswith("emb." ): snake_case_ = name.replace("emb." , "embeddings." ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("blocks.0.ln0" ): snake_case_ = name.replace("blocks.0.ln0" , "blocks.0.pre_ln" ) # att -> attention snake_case_ = re.sub(r"blocks\.(\d+)\.att" , r"blocks.\1.attention" , snake_case ) # ffn -> feed_forward snake_case_ = re.sub(r"blocks\.(\d+)\.ffn" , r"blocks.\1.feed_forward" , snake_case ) # time_mix_k -> time_mix_key and reshape if name.endswith(".time_mix_k" ): snake_case_ = name.replace(".time_mix_k" , ".time_mix_key" ) # time_mix_v -> time_mix_value and reshape if name.endswith(".time_mix_v" ): snake_case_ = name.replace(".time_mix_v" , ".time_mix_value" ) # time_mix_r -> time_mix_key and reshape if name.endswith(".time_mix_r" ): snake_case_ = name.replace(".time_mix_r" , ".time_mix_receptance" ) if name != "head.weight": snake_case_ = "rwkv." + name snake_case_ = weight return state_dict def UpperCamelCase_( snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : str=None , snake_case : Union[str, Any]=None , snake_case : Any=False , snake_case : Tuple=None ): '''simple docstring''' if tokenizer_file is None: print("No `--tokenizer_file` provided, we will use the default tokenizer." ) snake_case_ = 5_0_2_7_7 snake_case_ = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-20b" ) else: snake_case_ = PreTrainedTokenizerFast(tokenizer_file=snake_case ) snake_case_ = len(snake_case ) tokenizer.save_pretrained(snake_case ) # 2. Build the config snake_case_ = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: snake_case_ = candidate break if size is None: raise ValueError("Could not infer the size, please provide it with the `--size` argument." ) if size not in possible_sizes: raise ValueError(f'`size` should be one of {possible_sizes}, got {size}.' ) snake_case_ = RwkvConfig( vocab_size=snake_case , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(snake_case ) # 3. Download model file then convert state_dict snake_case_ = hf_hub_download(snake_case , snake_case ) snake_case_ = torch.load(snake_case , map_location="cpu" ) snake_case_ = convert_state_dict(snake_case ) # 4. Split in shards and save snake_case_ , snake_case_ = shard_checkpoint(snake_case ) for shard_file, shard in shards.items(): torch.save(snake_case , os.path.join(snake_case , snake_case ) ) if index is not None: snake_case_ = os.path.join(snake_case , snake_case ) # Save the index as well with open(snake_case , "w" , encoding="utf-8" ) as f: snake_case_ = json.dumps(snake_case , indent=2 , sort_keys=snake_case ) + "\n" f.write(snake_case ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( "Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model." ) snake_case_ = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: snake_case_ = torch.load(os.path.join(snake_case , snake_case ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(snake_case , snake_case ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("Please provide a `model_name` to push the model to the Hub." ) snake_case_ = AutoModelForCausalLM.from_pretrained(snake_case ) model.push_to_hub(snake_case , max_shard_size="2GB" ) tokenizer.push_to_hub(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--repo_id", default=None, type=str, required=True, help="Repo ID from which to pull the checkpoint." ) parser.add_argument( "--checkpoint_file", default=None, type=str, required=True, help="Name of the checkpoint file in the repo." ) parser.add_argument( "--output_dir", default=None, type=str, required=True, help="Where to save the converted model." ) parser.add_argument( "--tokenizer_file", default=None, type=str, help="Path to the tokenizer file to use (if not provided, only the model is converted).", ) parser.add_argument( "--size", default=None, type=str, help="Size of the model. Will be inferred from the `checkpoint_file` if not passed.", ) parser.add_argument( "--push_to_hub", action="store_true", help="Push to the Hub the converted model.", ) parser.add_argument( "--model_name", default=None, type=str, help="Name of the pushed model on the Hub, including the username / organization.", ) _SCREAMING_SNAKE_CASE : Dict = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
85
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> None: __A : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a ) == len(a ), F"""{len(a )} != {len(a )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) UpperCAmelCase : List[Any] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } UpperCAmelCase : Optional[int] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _SCREAMING_SNAKE_CASE ( a , a ) -> Dict: try: __A : int = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a ) ) def _SCREAMING_SNAKE_CASE ( a , a ) -> List[int]: if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _SCREAMING_SNAKE_CASE ( a , a = "student" , a = None , a = None , a=False , a=None , a=None , **a , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __A : List[str] = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(a , a ): AutoTokenizer.from_pretrained(a ).save_pretrained(a ) # purely for convenience __A : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained(a ).eval() else: assert isinstance(a , a ), F"""teacher must be a model or string got type {type(a )}""" __A : int = teacher.config.to_diff_dict() try: __A , __A : List[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A : str = teacher_e if d is None: __A : List[Any] = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): __A , __A : List[Any] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A : Optional[int] = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A : int = teacher_e if d is None: __A : Optional[Any] = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a ) # Copy weights __A : Dict = teacher.config_class(**a ) __A : int = AutoModelForSeqaSeqLM.from_config(a ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A : Any = student.load_state_dict(teacher.state_dict() , strict=a ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A : Optional[int] = list(range(a ) ), list(range(a ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) if d_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) try: if hasattr( a , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a ) copy_layers(teacher.decoder.block , student.decoder.block , a ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __A : Optional[int] = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(a ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
280
0
"""simple docstring""" # A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Any = [False] * len(_UpperCamelCase ) __lowerCAmelCase : int = [-1] * len(_UpperCamelCase ) def dfs(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Optional[int] = True __lowerCAmelCase : Any = c for u in graph[v]: if not visited[u]: dfs(_UpperCamelCase , 1 - c ) for i in range(len(_UpperCamelCase ) ): if not visited[i]: dfs(_UpperCamelCase , 0 ) for i in range(len(_UpperCamelCase ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph lowerCamelCase__ = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
86
def _SCREAMING_SNAKE_CASE ( a , a ) -> list[int]: __A : Optional[int] = int(a ) # Initialize Result __A : Optional[int] = [] # Traverse through all denomination for denomination in reversed(a ): # Find denominations while int(a ) >= int(a ): total_value -= int(a ) answer.append(a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase : List[str] = [] UpperCAmelCase : Optional[int] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): UpperCAmelCase : List[Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) UpperCAmelCase : int = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCAmelCase : Tuple = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(F"""Following is minimal change for {value}: """) UpperCAmelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
280
0
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def lowercase_ ( _lowerCamelCase : int): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set()) @pytest.fixture def lowercase_ ( _lowerCamelCase : Union[str, Any]): class snake_case_ : def __init__( self : int , lowercase_ : Dict ) -> Tuple: lowercase__ : List[Any] = metric_id class snake_case_ : __A : Any = [MetricMock(__A ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def __UpperCamelCase ( self : Tuple ) -> Tuple: return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock()) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))]) def lowercase_ ( _lowerCamelCase : List[Any] , _lowerCamelCase : int , _lowerCamelCase : str , _lowerCamelCase : Tuple , _lowerCamelCase : Dict): if "tmp_path" in args: lowercase__ : Any = tuple(arg if arg != "tmp_path" else tmp_path for arg in args) with pytest.warns(_lowerCamelCase , match="https://huggingface.co/docs/evaluate"): func(*_lowerCamelCase)
87
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _A( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __A : List[Any] = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __A : Union[str, Any] = parent __A : Optional[int] = batch_size __A : int = num_channels __A : int = min_resolution __A : Any = max_resolution __A : List[Any] = do_resize __A : List[Any] = size __A : Union[str, Any] = do_normalize __A : Optional[int] = image_mean __A : Optional[int] = image_std __A : int = do_rescale __A : str = rescale_factor __A : Tuple = do_pad def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase_ ( self , _A , _A=False ): if not batched: __A : List[str] = image_inputs[0] if isinstance(_A , Image.Image ): __A , __A : int = image.size else: __A , __A : Any = image.shape[1], image.shape[2] if w < h: __A : List[Any] = int(self.size['shortest_edge'] * h / w ) __A : List[Any] = self.size['shortest_edge'] elif w > h: __A : Union[str, Any] = self.size['shortest_edge'] __A : str = int(self.size['shortest_edge'] * w / h ) else: __A : Dict = self.size['shortest_edge'] __A : str = self.size['shortest_edge'] else: __A : int = [] for image in image_inputs: __A , __A : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __A : List[str] = max(_A , key=lambda _A : item[0] )[0] __A : str = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : List[str] = YolosImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): __A : Dict = YolosImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): __A : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) def UpperCAmelCase_ ( self ): __A : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , _A ) __A : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _A ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing __A : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input __A : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A , __A : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) __A : str = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input __A : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : List[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Tuple = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input __A : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Union[str, Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Optional[int] = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processings __A : Tuple = self.image_processing_class(**self.image_processor_dict ) __A : Any = self.image_processing_class(do_resize=_A , do_normalize=_A , do_rescale=_A ) # create random PyTorch tensors __A : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __A : Optional[int] = image_processing_a.pad(_A , return_tensors='pt' ) __A : Optional[int] = image_processing_a(_A , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1e-4 ) ) @slow def UpperCAmelCase_ ( self ): # prepare image and target __A : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __A : Optional[Any] = json.loads(f.read() ) __A : Optional[Any] = {'image_id': 39769, 'annotations': target} # encode them __A : str = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) __A : List[Any] = image_processing(images=_A , annotations=_A , return_tensors='pt' ) # verify pixel values __A : List[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : List[Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Any = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify orig_size __A : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : str = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) ) @slow def UpperCAmelCase_ ( self ): # prepare image, target and masks_path __A : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __A : Tuple = json.loads(f.read() ) __A : Any = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __A : List[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __A : Any = YolosImageProcessor(format='coco_panoptic' ) __A : List[Any] = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='pt' ) # verify pixel values __A : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Union[str, Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : List[str] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify masks __A : Tuple = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _A ) # verify orig_size __A : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : int = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) )
280
0
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __lowerCAmelCase : int = random.Random() if is_torch_available(): import torch def a__ ( A_, A_=1.0, A_=None, A_=None ): '''simple docstring''' if rng is None: __magic_name__ = global_rng __magic_name__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str=7 , UpperCamelCase__ : Union[str, Any]=400 , UpperCamelCase__ : List[str]=2000 , UpperCamelCase__ : Optional[int]=1 , UpperCamelCase__ : List[str]=0.0 , UpperCamelCase__ : Optional[int]=1_6000 , UpperCamelCase__ : str=True , UpperCamelCase__ : Tuple=True , ) -> Any: """simple docstring""" __magic_name__ = parent __magic_name__ = batch_size __magic_name__ = min_seq_length __magic_name__ = max_seq_length __magic_name__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __magic_name__ = feature_size __magic_name__ = padding_value __magic_name__ = sampling_rate __magic_name__ = return_attention_mask __magic_name__ = do_normalize def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowercase ( self : Dict , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : Dict=False ) -> List[str]: """simple docstring""" def _flatten(UpperCamelCase__ : List[str] ): return list(itertools.chain(*UpperCamelCase__ ) ) if equal_length: __magic_name__ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __magic_name__ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __magic_name__ = [np.asarray(UpperCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class UpperCAmelCase_ ( _A , unittest.TestCase ): '''simple docstring''' a__ = ASTFeatureExtractor def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" __magic_name__ = ASTFeatureExtractionTester(self ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" __magic_name__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __magic_name__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ = [np.asarray(UpperCamelCase__ ) for speech_input in speech_inputs] # Test not batched input __magic_name__ = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values __magic_name__ = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) ) # Test batched __magic_name__ = feat_extract(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors="""np""" ).input_values __magic_name__ = feat_extract(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertTrue(np.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __magic_name__ = [floats_list((1, x) )[0] for x in (800, 800, 800)] __magic_name__ = np.asarray(UpperCamelCase__ ) __magic_name__ = feat_extract(UpperCamelCase__ , return_tensors="""np""" ).input_values __magic_name__ = feat_extract(UpperCamelCase__ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertTrue(np.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) ) @require_torch def _lowercase ( self : Dict ) -> int: """simple docstring""" import torch __magic_name__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ = np.random.rand(100 ).astype(np.floataa ) __magic_name__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __magic_name__ = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __magic_name__ = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def _lowercase ( self : Tuple , UpperCamelCase__ : str ) -> Tuple: """simple docstring""" from datasets import load_dataset __magic_name__ = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __magic_name__ = ds.sort("""id""" ).select(range(UpperCamelCase__ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] @require_torch def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" __magic_name__ = torch.tensor( [-0.9894, -1.2776, -0.9066, -1.2776, -0.9349, -1.2609, -1.0386, -1.2776, -1.1561, -1.2776, -1.2052, -1.2723, -1.2190, -1.2132, -1.2776, -1.1133, -1.1953, -1.1343, -1.1584, -1.2203, -1.1770, -1.2474, -1.2381, -1.1936, -0.9270, -0.8317, -0.8049, -0.7706, -0.7565, -0.7869] ) # fmt: on __magic_name__ = self._load_datasamples(1 ) __magic_name__ = ASTFeatureExtractor() __magic_name__ = feature_extractor(UpperCamelCase__ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 1024, 128) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , UpperCamelCase__ , atol=1E-4 ) )
88
import argparse import json from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( ) -> List[Any]: __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=a , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=a , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=a , help='where to store parsed gold_data_path file' , ) __A : Optional[int] = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: __A : List[Any] = json.load(a ) for dpr_record in tqdm(a ): __A : Dict = dpr_record['question'] __A : Any = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(a ) + '\n' ) if __name__ == "__main__": main()
280
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = {'''configuration_beit''': ['''BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BeitConfig''', '''BeitOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['''BeitFeatureExtractor'''] __lowerCAmelCase = ['''BeitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''BEIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BeitForImageClassification''', '''BeitForMaskedImageModeling''', '''BeitForSemanticSegmentation''', '''BeitModel''', '''BeitPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''FlaxBeitForImageClassification''', '''FlaxBeitForMaskedImageModeling''', '''FlaxBeitModel''', '''FlaxBeitPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
89
from heapq import heappop, heappush import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a , ) -> tuple[float | int, list[tuple[int, int]]]: __A , __A : int = grid.shape __A : Any = [-1, 1, 0, 0] __A : Optional[Any] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __A , __A : Optional[int] = [(0, source)], set() __A : Any = np.full((rows, cols) , np.inf ) __A : Any = 0 __A : Any = np.empty((rows, cols) , dtype=a ) __A : Optional[Any] = None while queue: ((__A) , (__A)) : List[str] = heappop(a ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __A : int = [] while (x, y) != source: path.append((x, y) ) __A , __A : Optional[int] = predecessors[x, y] path.append(a ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a ) ): __A , __A : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __A : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a , (dist + 1, (nx, ny)) ) __A : List[Any] = dist + 1 __A : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
280
0
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore __A = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" __A = [file for file in filepaths if file != file.lower()] if upper_files: print(f'''{len(upper_files)} files contain uppercase characters:''') print("\n".join(upper_files) + "\n") __A = [file for file in filepaths if " " in file] if space_files: print(f'''{len(space_files)} files contain space characters:''') print("\n".join(space_files) + "\n") __A = [file for file in filepaths if "-" in file] if hyphen_files: print(f'''{len(hyphen_files)} files contain hyphen characters:''') print("\n".join(hyphen_files) + "\n") __A = [file for file in filepaths if os.sep not in file] if nodir_files: print(f'''{len(nodir_files)} files are not in a directory:''') print("\n".join(nodir_files) + "\n") __A = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
90
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
0
"""simple docstring""" from pathlib import Path import numpy as np from PIL import Image def _A (__a ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[Any] = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def _A (__a ) -> np.ndarray: """simple docstring""" return (gray > 1_27) & (gray <= 2_55) def _A (__a , __a ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE_ : int = np.zeros_like(__a ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image SCREAMING_SNAKE_CASE_ : Union[str, Any] = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): SCREAMING_SNAKE_CASE_ : int = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() SCREAMING_SNAKE_CASE_ : Any = int(summation > 0 ) return output if __name__ == "__main__": # read original image UpperCAmelCase_ : Dict = Path(__file__).resolve().parent / """image_data""" / """lena.jpg""" UpperCAmelCase_ : int = np.array(Image.open(lena_path)) # kernel to be applied UpperCAmelCase_ : Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) UpperCAmelCase_ : Union[str, Any] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image UpperCAmelCase_ : List[Any] = Image.fromarray(output).convert("""RGB""") pil_img.save("""result_dilation.png""")
91
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _SCREAMING_SNAKE_CASE ( a , a ) -> Optional[Any]: __A : Any = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 10_24, 'hidden_size': 7_68, 'max_length': 5_12, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 10_24, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } __A : str = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __A : Optional[int] = BERTEncoder( attention_cell=predefined_args['attention_cell'] , num_layers=predefined_args['num_layers'] , units=predefined_args['units'] , hidden_size=predefined_args['hidden_size'] , max_length=predefined_args['max_length'] , num_heads=predefined_args['num_heads'] , scaled=predefined_args['scaled'] , dropout=predefined_args['dropout'] , output_attention=a , output_all_encodings=a , use_residual=predefined_args['use_residual'] , activation=predefined_args.get('activation' , 'gelu' ) , layer_norm_eps=predefined_args.get('layer_norm_eps' , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __A : Union[str, Any] = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab __A : Any = os.path.join(get_home_dir() , 'models' ) __A : List[Any] = _load_vocab(a , a , a , cls=a ) __A : Dict = nlp.model.BERTModel( a , len(a ) , units=predefined_args['units'] , embed_size=predefined_args['embed_size'] , embed_dropout=predefined_args['embed_dropout'] , word_embed=predefined_args['word_embed'] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args['token_type_vocab_size'] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) __A : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __A : Any = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(a ), } __A : int = BertConfig.from_dict(a ) __A : Union[str, Any] = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a , a ): __A : Tuple = hf_param.shape __A : str = to_torch(params[gluon_param] ) __A : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), F"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __A : str = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , 'word_embed.0.weight' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , 'encoder.position_weight' ) __A : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , 'encoder.layer_norm.beta' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , 'encoder.layer_norm.gamma' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __A : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __A : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __A : BertSelfAttention = layer.attention.self __A : Optional[Any] = check_and_map_params( self_attn.key.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.key.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.query.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __A : Optional[Any] = check_and_map_params( self_attn.query.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.value.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.value.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __A : BertSelfOutput = layer.attention.output __A : Tuple = check_and_map_params( self_output.dense.bias , F"""encoder.transformer_cells.{i}.proj.bias""" ) __A : int = check_and_map_params( self_output.dense.weight , F"""encoder.transformer_cells.{i}.proj.weight""" ) __A : List[Any] = check_and_map_params( self_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __A : str = check_and_map_params( self_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __A : BertIntermediate = layer.intermediate __A : int = check_and_map_params( intermediate.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __A : List[Any] = check_and_map_params( intermediate.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __A : BertOutput = layer.output __A : List[Any] = check_and_map_params( bert_output.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __A : Dict = check_and_map_params( bert_output.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __A : Optional[int] = check_and_map_params( bert_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __A : Dict = check_and_map_params( bert_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __A : Any = RobertaTokenizer.from_pretrained('roberta-base' ) __A : List[str] = tokenizer.encode_plus(a )['input_ids'] # Get gluon output __A : List[str] = mx.nd.array([input_ids] ) __A : Union[str, Any] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) __A : Optional[Any] = BertModel.from_pretrained(a ) hf_bort_model.eval() __A : Tuple = tokenizer.encode_plus(a , return_tensors='pt' ) __A : Any = hf_bort_model(**a )[0] __A : Union[str, Any] = output_gluon[0].asnumpy() __A : Tuple = output_hf[0].detach().numpy() __A : int = np.max(np.abs(hf_layer - gluon_layer ) ).item() __A : int = np.allclose(a , a , atol=1e-3 ) if success: print('✔️ Both model do output the same tensors' ) else: print('❌ Both model do **NOT** output the same tensors' ) print('Absolute difference is:' , a ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase : Dict = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
280
0
import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class a__ ( unittest.TestCase ): def __init__( self , _A , _A=1_3 , _A=7 , _A=True , _A=True , _A=True , _A=True , _A=9_9 , _A=3_2 , _A=5 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_2 , _A=1_6 , _A=2 , _A=0.02 , _A=4 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_attention_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_choices def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_attention_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_A , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = True __lowerCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a__ ( snake_case__ , unittest.TestCase ): _a : List[str] = True _a : List[str] = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxBertModelTester(self ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxBertModel.from_pretrained("bert-base-cased" ) __lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_A )
92
import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: __A : List[str] = x __A : str = y for step in range(a ): # noqa: B007 __A : Union[str, Any] = a * a - b * b + x __A : Optional[int] = 2 * a * b + y __A : List[str] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( a = 8_00 , a = 6_00 , a = -0.6 , a = 0 , a = 3.2 , a = 50 , a = True , ) -> Image.Image: __A : str = Image.new('RGB' , (image_width, image_height) ) __A : Dict = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates __A : Dict = figure_width / image_width * image_height __A : Union[str, Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width __A : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height __A : Union[str, Any] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __A : Optional[Any] = get_color_coded_rgb(a ) else: __A : Dict = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCAmelCase : str = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
280
0
'''simple docstring''' import warnings from .generation import TFGenerationMixin class lowerCAmelCase__ ( lowerCamelCase_ ): # warning at import time warnings.warn( '''Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ''' '''be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.''' , lowerCamelCase_ , )
93
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( a , nominal_annual_percentage_rate / 3_65 , number_of_years * 3_65 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
94
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
import math def _A ( SCREAMING_SNAKE_CASE : int ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _A ( SCREAMING_SNAKE_CASE : int = 10_001 ): """simple docstring""" try: a__ : Optional[int] =int(SCREAMING_SNAKE_CASE ) except (TypeError, ValueError): raise TypeError("Parameter nth must be int or castable to int." ) from None if nth <= 0: raise ValueError("Parameter nth must be greater than or equal to one." ) a__ : list[int] =[] a__ : int =2 while len(SCREAMING_SNAKE_CASE ) < nth: if is_prime(SCREAMING_SNAKE_CASE ): primes.append(SCREAMING_SNAKE_CASE ) num += 1 else: num += 1 return primes[len(SCREAMING_SNAKE_CASE ) - 1] if __name__ == "__main__": print(F"""{solution() = }""")
95
def _SCREAMING_SNAKE_CASE ( a ) -> bool: return str(a ) == str(a )[::-1] def _SCREAMING_SNAKE_CASE ( a ) -> int: return int(a ) + int(str(a )[::-1] ) def _SCREAMING_SNAKE_CASE ( a = 1_00_00 ) -> int: __A : int = [] for num in range(1 , a ): __A : List[str] = 0 __A : List[Any] = num while iterations < 50: __A : str = sum_reverse(a ) iterations += 1 if is_palindrome(a ): break else: lychrel_nums.append(a ) return len(a ) if __name__ == "__main__": print(F"""{solution() = }""")
280
0
"""simple docstring""" from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """t5-small""": """https://huggingface.co/t5-small/resolve/main/config.json""", """t5-base""": """https://huggingface.co/t5-base/resolve/main/config.json""", """t5-large""": """https://huggingface.co/t5-large/resolve/main/config.json""", """t5-3b""": """https://huggingface.co/t5-3b/resolve/main/config.json""", """t5-11b""": """https://huggingface.co/t5-11b/resolve/main/config.json""", } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """t5""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , lowercase=32128 , lowercase=512 , lowercase=64 , lowercase=2048 , lowercase=6 , lowercase=None , lowercase=8 , lowercase=32 , lowercase=128 , lowercase=0.1 , lowercase=1E-6 , lowercase=1.0 , lowercase="relu" , lowercase=True , lowercase=True , lowercase=0 , lowercase=1 , **lowercase , ): _lowerCamelCase : Optional[int] = vocab_size _lowerCamelCase : Any = d_model _lowerCamelCase : List[str] = d_kv _lowerCamelCase : Optional[int] = d_ff _lowerCamelCase : Union[str, Any] = num_layers _lowerCamelCase : str = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _lowerCamelCase : Union[str, Any] = num_heads _lowerCamelCase : Any = relative_attention_num_buckets _lowerCamelCase : Tuple = relative_attention_max_distance _lowerCamelCase : List[str] = dropout_rate _lowerCamelCase : Optional[Any] = layer_norm_epsilon _lowerCamelCase : Dict = initializer_factor _lowerCamelCase : Dict = feed_forward_proj _lowerCamelCase : Union[str, Any] = use_cache _lowerCamelCase : Tuple = self.feed_forward_proj.split('-' ) _lowerCamelCase : Tuple = act_info[-1] _lowerCamelCase : str = act_info[0] == 'gated' if len(lowercase ) > 1 and act_info[0] != "gated" or len(lowercase ) > 2: raise ValueError( F'''`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.''' 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": _lowerCamelCase : int = 'gelu_new' super().__init__( pad_token_id=lowercase , eos_token_id=lowercase , is_encoder_decoder=lowercase , **lowercase , ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' @property def A_ ( self ): _lowerCamelCase : Any = { 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: _lowerCamelCase : int = 'past_encoder_sequence + sequence' _lowerCamelCase : Union[str, Any] = {0: 'batch'} _lowerCamelCase : Optional[Any] = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: _lowerCamelCase : Tuple = {0: 'batch', 1: 'decoder_sequence'} _lowerCamelCase : Dict = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(lowercase , direction='inputs' ) return common_inputs @property def A_ ( self ): return 13
96
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class _A: """simple docstring""" def __init__( self , _A = None ): if components is None: __A : int = [] __A : Tuple = list(_A ) def __len__( self ): return len(self.__components ) def __str__( self ): return "(" + ",".join(map(_A , self.__components ) ) + ")" def __add__( self , _A ): __A : Optional[int] = len(self ) if size == len(_A ): __A : Any = [self.__components[i] + other.component(_A ) for i in range(_A )] return Vector(_A ) else: raise Exception('must have the same size' ) def __sub__( self , _A ): __A : Tuple = len(self ) if size == len(_A ): __A : Union[str, Any] = [self.__components[i] - other.component(_A ) for i in range(_A )] return Vector(_A ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , (float, int) ): __A : str = [c * other for c in self.__components] return Vector(_A ) elif isinstance(_A , _A ) and len(self ) == len(_A ): __A : Union[str, Any] = len(self ) __A : Dict = [self.__components[i] * other.component(_A ) for i in range(_A )] return sum(_A ) else: # error case raise Exception('invalid operand!' ) def UpperCAmelCase_ ( self ): return Vector(self.__components ) def UpperCAmelCase_ ( self , _A ): if isinstance(_A , _A ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def UpperCAmelCase_ ( self , _A , _A ): assert -len(self.__components ) <= pos < len(self.__components ) __A : Optional[int] = value def UpperCAmelCase_ ( self ): if len(self.__components ) == 0: raise Exception('Vector is empty' ) __A : Optional[Any] = [c**2 for c in self.__components] return math.sqrt(sum(_A ) ) def UpperCAmelCase_ ( self , _A , _A = False ): __A : Optional[Any] = self * other __A : Optional[Any] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def _SCREAMING_SNAKE_CASE ( a ) -> Vector: assert isinstance(a , a ) return Vector([0] * dimension ) def _SCREAMING_SNAKE_CASE ( a , a ) -> Vector: assert isinstance(a , a ) and (isinstance(a , a )) __A : Optional[Any] = [0] * dimension __A : Tuple = 1 return Vector(a ) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: assert ( isinstance(a , a ) and isinstance(a , a ) and (isinstance(a , (int, float) )) ) return x * scalar + y def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: random.seed(a ) __A : str = [random.randint(a , a ) for _ in range(a )] return Vector(a ) class _A: """simple docstring""" def __init__( self , _A , _A , _A ): __A : Optional[Any] = matrix __A : Dict = w __A : Optional[int] = h def __str__( self ): __A : Tuple = '' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Optional[Any] = [] for i in range(self.__height ): __A : Optional[Any] = [ self.__matrix[i][j] + other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Tuple = [] for i in range(self.__height ): __A : str = [ self.__matrix[i][j] - other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , _A ): # matrix-vector if len(_A ) == self.__width: __A : List[Any] = zero_vector(self.__height ) for i in range(self.__height ): __A : List[str] = [ self.__matrix[i][j] * other.component(_A ) for j in range(self.__width ) ] ans.change_component(_A , sum(_A ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(_A , (int, float) ): # matrix-scalar __A : List[str] = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(_A , self.__width , self.__height ) return None def UpperCAmelCase_ ( self ): return self.__height def UpperCAmelCase_ ( self ): return self.__width def UpperCAmelCase_ ( self , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: __A : int = value else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) __A : List[str] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(_A ) ): __A : Optional[int] = minor[i][:y] + minor[i][y + 1 :] return Matrix(_A , self.__width - 1 , self.__height - 1 ).determinant() def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(_A , _A ) else: raise Exception('Indices out of bounds' ) def UpperCAmelCase_ ( self ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __A : List[str] = [ self.__matrix[0][y] * self.cofactor(0 , _A ) for y in range(self.__width ) ] return sum(_A ) def _SCREAMING_SNAKE_CASE ( a ) -> Matrix: __A : list[list[float]] = [[0] * n for _ in range(a )] return Matrix(a , a , a ) def _SCREAMING_SNAKE_CASE ( a , a , a , a ) -> Matrix: random.seed(a ) __A : list[list[float]] = [ [random.randint(a , a ) for _ in range(a )] for _ in range(a ) ] return Matrix(a , a , a )
280
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class lowercase ( A__ ): """simple docstring""" _a = 'microsoft/speecht5_tts' _a = ( 'This is a tool that reads an English text out loud. It takes an input named `text` which should contain the ' 'text to read (in English) and returns a waveform object containing the sound.' ) _a = 'text_reader' _a = SpeechTaProcessor _a = SpeechTaForTextToSpeech _a = SpeechTaHifiGan _a = ['text'] _a = ['audio'] def lowerCAmelCase__ ( self ): '''simple docstring''' if self.post_processor is None: UpperCamelCase__ :int = '''microsoft/speecht5_hifigan''' super().setup() def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_=None ): '''simple docstring''' UpperCamelCase__ :str = self.pre_processor(text=UpperCamelCase_ , return_tensors='''pt''' , truncation=UpperCamelCase_ ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError('''Datasets needs to be installed if not passing speaker embeddings.''' ) UpperCamelCase__ :Union[str, Any] = load_dataset('''Matthijs/cmu-arctic-xvectors''' , split='''validation''' ) UpperCamelCase__ :int = torch.tensor(embeddings_dataset[7305]['''xvector'''] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' with torch.no_grad(): return self.model.generate_speech(**UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' with torch.no_grad(): return self.post_processor(UpperCamelCase_ ).cpu().detach()
97
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase : List[str] = '''▁''' UpperCAmelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Optional[int] = BertGenerationTokenizer UpperCamelCase : str = False UpperCamelCase : Tuple = True def UpperCAmelCase_ ( self ): super().setUp() __A : Tuple = BertGenerationTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : str = '<s>' __A : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def UpperCAmelCase_ ( self ): __A : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(_A ) , 1002 ) def UpperCAmelCase_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCAmelCase_ ( self ): __A : str = BertGenerationTokenizer(_A , keep_accents=_A ) __A : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_A , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [285, 46, 10, 170, 382] , ) __A : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __A : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __A : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase_ ( self ): return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def UpperCAmelCase_ ( self ): __A : List[Any] = 'Hello World!' __A : Optional[Any] = [18536, 2260, 101] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def UpperCAmelCase_ ( self ): __A : Dict = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) __A : int = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @require_torch @slow def UpperCAmelCase_ ( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence __A : Tuple = list(self.big_tokenizer.get_vocab().keys() )[:10] __A : List[Any] = ' '.join(_A ) __A : Union[str, Any] = self.big_tokenizer.encode_plus(_A , return_tensors='pt' , return_token_type_ids=_A ) __A : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=_A ) __A : int = BertGenerationConfig() __A : List[str] = BertGenerationEncoder(_A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def UpperCAmelCase_ ( self ): # fmt: off __A : str = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
280
0
"""simple docstring""" import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) def a_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None ): # Recurse if needed if "." in tensor_name: UpperCAmelCase__ = tensor_name.split('.' ) for split in splits[:-1]: UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) if new_module is None: raise ValueError(f'''{module} has no attribute {split}.''' ) UpperCAmelCase__ = new_module UpperCAmelCase__ = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f'''{module} does not have a parameter or a buffer named {tensor_name}.''' ) UpperCAmelCase__ = tensor_name in module._buffers UpperCAmelCase__ = getattr(lowerCamelCase , lowerCamelCase ) if old_value.device == torch.device('meta' ) and device not in ["meta", torch.device('meta' )] and value is None: raise ValueError(f'''{tensor_name} is on the meta device, we need a `value` to put in on {device}.''' ) UpperCAmelCase__ = False UpperCAmelCase__ = False if is_buffer or not is_bitsandbytes_available(): UpperCAmelCase__ = False UpperCAmelCase__ = False else: UpperCAmelCase__ = hasattr(bnb.nn , 'Params4bit' ) and isinstance(module._parameters[tensor_name] , bnb.nn.Paramsabit ) UpperCAmelCase__ = isinstance(module._parameters[tensor_name] , bnb.nn.IntaParams ) if is_abit or is_abit: UpperCAmelCase__ = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: UpperCAmelCase__ = old_value.to(lowerCamelCase ) elif isinstance(lowerCamelCase , torch.Tensor ): UpperCAmelCase__ = value.to('cpu' ) if value.dtype == torch.inta: UpperCAmelCase__ = version.parse(importlib.metadata.version('bitsandbytes' ) ) > version.parse( '0.37.2' ) if not is_abit_serializable: raise ValueError( 'Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ' 'Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.' ) else: UpperCAmelCase__ = torch.tensor(lowerCamelCase , device='cpu' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls , lowerCamelCase ) and fpaa_statistics is None: UpperCAmelCase__ = new_value.T UpperCAmelCase__ = old_value.__dict__ if is_abit: UpperCAmelCase__ = bnb.nn.IntaParams(lowerCamelCase , requires_grad=lowerCamelCase , **lowerCamelCase ).to(lowerCamelCase ) elif is_abit: UpperCAmelCase__ = bnb.nn.Paramsabit(lowerCamelCase , requires_grad=lowerCamelCase , **lowerCamelCase ).to(lowerCamelCase ) UpperCAmelCase__ = new_value if fpaa_statistics is not None: setattr(module.weight , 'SCB' , fpaa_statistics.to(lowerCamelCase ) ) else: if value is None: UpperCAmelCase__ = old_value.to(lowerCamelCase ) elif isinstance(lowerCamelCase , torch.Tensor ): UpperCAmelCase__ = value.to(lowerCamelCase ) else: UpperCAmelCase__ = torch.tensor(lowerCamelCase , device=lowerCamelCase ) if is_buffer: UpperCAmelCase__ = new_value else: UpperCAmelCase__ = nn.Parameter(lowerCamelCase , requires_grad=old_value.requires_grad ) UpperCAmelCase__ = new_value def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=False ): for name, module in model.named_children(): if current_key_name is None: UpperCAmelCase__ = [] current_key_name.append(lowerCamelCase ) if (isinstance(lowerCamelCase , nn.Linear ) or isinstance(lowerCamelCase , lowerCamelCase )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '.'.join(lowerCamelCase ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ , UpperCAmelCase__ = module.weight.shape else: UpperCAmelCase__ = module.in_features UpperCAmelCase__ = module.out_features if quantization_config.quantization_method() == "llm_int8": UpperCAmelCase__ = bnb.nn.LinearabitLt( lowerCamelCase , lowerCamelCase , module.bias is not None , has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight , threshold=quantization_config.llm_inta_threshold , ) UpperCAmelCase__ = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: UpperCAmelCase__ = bnb.nn.Linearabit( lowerCamelCase , lowerCamelCase , module.bias is not None , quantization_config.bnb_abit_compute_dtype , compress_statistics=quantization_config.bnb_abit_use_double_quant , quant_type=quantization_config.bnb_abit_quant_type , ) UpperCAmelCase__ = True # Store the module class in case we need to transpose the weight later UpperCAmelCase__ = type(lowerCamelCase ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(lowerCamelCase ) if len(list(module.children() ) ) > 0: UpperCAmelCase__ , UpperCAmelCase__ = _replace_with_bnb_linear( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , has_been_replaced=lowerCamelCase , ) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a_ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None ): UpperCAmelCase__ = ['lm_head'] if modules_to_not_convert is None else modules_to_not_convert UpperCAmelCase__ , UpperCAmelCase__ = _replace_with_bnb_linear( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if not has_been_replaced: logger.warning( 'You are loading your model in 8bit or 4bit but no linear modules were found in your model.' ' Please double check your model architecture, or submit an issue on github if you think this is' ' a bug.' ) return model def a_ ( *lowerCamelCase , **lowerCamelCase ): warnings.warn( '`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead' , lowerCamelCase , ) return replace_with_bnb_linear(*lowerCamelCase , **lowerCamelCase ) def a_ ( *lowerCamelCase , **lowerCamelCase ): warnings.warn( '`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead' , lowerCamelCase , ) return set_module_quantized_tensor_to_device(*lowerCamelCase , **lowerCamelCase ) def a_ ( lowerCamelCase ): UpperCAmelCase__ = deepcopy(lowerCamelCase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() UpperCAmelCase__ = find_tied_parameters(lowerCamelCase ) # For compatibility with Accelerate < 0.18 if isinstance(lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: UpperCAmelCase__ = sum(lowerCamelCase , [] ) UpperCAmelCase__ = len(lowerCamelCase ) > 0 # Check if it is a base model UpperCAmelCase__ = not hasattr(lowerCamelCase , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head UpperCAmelCase__ = list(model.named_children() ) UpperCAmelCase__ = [list_modules[-1][0]] # add last module together with tied weights UpperCAmelCase__ = set(lowerCamelCase ) - set(lowerCamelCase ) UpperCAmelCase__ = list(set(lowerCamelCase ) ) + list(lowerCamelCase ) # remove ".weight" from the keys UpperCAmelCase__ = ['.weight', '.bias'] UpperCAmelCase__ = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: UpperCAmelCase__ = name.replace(lowerCamelCase , '' ) filtered_module_names.append(lowerCamelCase ) return filtered_module_names
98
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _A: """simple docstring""" @staticmethod def UpperCAmelCase_ ( *_A , **_A ): pass def _SCREAMING_SNAKE_CASE ( a ) -> str: __A : str = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def _SCREAMING_SNAKE_CASE ( a ) -> Dict: __A : Dict = np.array(a ) __A : List[Any] = npimg.shape return {"hash": hashimage(a ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _A( unittest.TestCase ): """simple docstring""" UpperCamelCase : str = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase : int = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def UpperCAmelCase_ ( self , _A , _A , _A ): __A : Dict = MaskGenerationPipeline(model=_A , image_processor=_A ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase_ ( self , _A , _A ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def UpperCAmelCase_ ( self ): pass @slow @require_torch def UpperCAmelCase_ ( self ): __A : Union[str, Any] = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) __A : List[str] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=256 ) # Shortening by hashing __A : List[Any] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] , ) # fmt: on @require_torch @slow def UpperCAmelCase_ ( self ): __A : Optional[Any] = 'facebook/sam-vit-huge' __A : List[str] = pipeline('mask-generation' , model=_A ) __A : Tuple = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __A : List[str] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] , )
280
0
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata lowercase : Tuple = """""" if version.parse(importlib_metadata.version("""jiwer""")) < version.parse("""2.3.0"""): class A__ ( tr.AbstractTransform ): """simple docstring""" def __init__( self , lowercase = " ") -> Tuple: '''simple docstring''' a__ : Tuple = sentence_delimiter def __lowercase ( self , lowercase) -> Optional[int]: '''simple docstring''' return list(lowercase) def __lowercase ( self , lowercase) -> Dict: '''simple docstring''' a__ : Tuple = [] for sent_idx, sentence in enumerate(lowercase): chars.extend(self.process_string(lowercase)) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(lowercase) - 1: chars.append(self.sentence_delimiter) return chars lowercase : Union[str, Any] = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: lowercase : List[str] = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) lowercase : List[Any] = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ lowercase : Optional[int] = """\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER's output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. """ lowercase : Optional[Any] = """ Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> cer = datasets.load_metric(\"cer\") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): """simple docstring""" def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence'), 'references': datasets.Value('string' , id='sequence'), }) , codebase_urls=['https://github.com/jitsi/jiwer/'] , reference_urls=[ 'https://en.wikipedia.org/wiki/Word_error_rate', 'https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates', ] , ) def __lowercase ( self , lowercase , lowercase , lowercase=False) -> Any: '''simple docstring''' if concatenate_texts: return jiwer.compute_measures( lowercase , lowercase , truth_transform=lowercase , hypothesis_transform=lowercase , )["wer"] a__ : Optional[int] = 0 a__ : str = 0 for prediction, reference in zip(lowercase , lowercase): a__ : Optional[int] = jiwer.compute_measures( lowercase , lowercase , truth_transform=lowercase , hypothesis_transform=lowercase , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
99
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): __A : List[Any] = tempfile.mkdtemp() # fmt: off __A : List[str] = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on __A : Union[str, Any] = dict(zip(_A , range(len(_A ) ) ) ) __A : Optional[int] = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] __A : int = {'unk_token': '<unk>'} __A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_A ) ) __A : List[Any] = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], 'image_std': [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } __A : Optional[int] = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def UpperCAmelCase_ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __A : Optional[int] = [Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase_ ( self ): __A : List[Any] = self.get_tokenizer() __A : str = self.get_rust_tokenizer() __A : List[str] = self.get_image_processor() __A : Optional[int] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_slow.save_pretrained(self.tmpdirname ) __A : int = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_fast.save_pretrained(self.tmpdirname ) __A : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _A ) self.assertIsInstance(processor_fast.tokenizer , _A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _A ) self.assertIsInstance(processor_fast.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : List[str] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) __A : Optional[int] = self.get_image_processor(do_normalize=_A ) __A : Any = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Optional[Any] = self.get_tokenizer() __A : Union[str, Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Union[str, Any] = self.prepare_image_inputs() __A : int = image_processor(_A , return_tensors='np' ) __A : str = processor(images=_A , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCAmelCase_ ( self ): __A : str = self.get_image_processor() __A : str = self.get_tokenizer() __A : Tuple = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : str = 'lower newer' __A : str = processor(text=_A , return_tensors='np' ) __A : List[str] = tokenizer(_A , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCAmelCase_ ( self ): __A : int = self.get_image_processor() __A : Optional[int] = self.get_tokenizer() __A : List[str] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Any = 'lower newer' __A : Optional[Any] = self.prepare_image_inputs() __A : List[Any] = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Any = 'google/owlvit-base-patch32' __A : int = OwlViTProcessor.from_pretrained(_A ) __A : Dict = ['cat', 'nasa badge'] __A : Optional[Any] = processor(text=_A ) __A : Optional[int] = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Tuple = 'google/owlvit-base-patch32' __A : Any = OwlViTProcessor.from_pretrained(_A ) __A : Dict = [['cat', 'nasa badge'], ['person']] __A : Dict = processor(text=_A ) __A : Optional[int] = 16 __A : Any = len(_A ) __A : Union[str, Any] = max([len(_A ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : List[Any] = 'google/owlvit-base-patch32' __A : str = OwlViTProcessor.from_pretrained(_A ) __A : Union[str, Any] = ['cat', 'nasa badge'] __A : Tuple = processor(text=_A ) __A : str = 16 __A : int = inputs['input_ids'] __A : List[Any] = [ [49406, 2368, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49406, 6841, 11301, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : List[str] = self.get_tokenizer() __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = processor(images=_A , query_images=_A ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Union[str, Any] = self.get_tokenizer() __A : str = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A : Any = processor.batch_decode(_A ) __A : Tuple = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A )
280
0
"""simple docstring""" import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("9.1.0"): __magic_name__ = { "linear": PIL.Image.Resampling.BILINEAR, "bilinear": PIL.Image.Resampling.BILINEAR, "bicubic": PIL.Image.Resampling.BICUBIC, "lanczos": PIL.Image.Resampling.LANCZOS, "nearest": PIL.Image.Resampling.NEAREST, } else: __magic_name__ = { "linear": PIL.Image.LINEAR, "bilinear": PIL.Image.BILINEAR, "bicubic": PIL.Image.BICUBIC, "lanczos": PIL.Image.LANCZOS, "nearest": PIL.Image.NEAREST, } def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = (images / 2 + 0.5).clamp(0 , 1 ) __SCREAMING_SNAKE_CASE = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __SCREAMING_SNAKE_CASE = numpy_to_pil(UpperCamelCase_ ) return images def _lowerCAmelCase ( UpperCamelCase_ ): if images.ndim == 3: __SCREAMING_SNAKE_CASE = images[None, ...] __SCREAMING_SNAKE_CASE = (images * 255).round().astype("""uint8""" ) if images.shape[-1] == 1: # special case for grayscale (single channel) images __SCREAMING_SNAKE_CASE = [Image.fromarray(image.squeeze() , mode="""L""" ) for image in images] else: __SCREAMING_SNAKE_CASE = [Image.fromarray(UpperCamelCase_ ) for image in images] return pil_images
100
import math def _SCREAMING_SNAKE_CASE ( a ) -> list[int]: __A : List[str] = [] __A : Any = 2 __A : Union[str, Any] = int(math.sqrt(a ) ) # Size of every segment __A : Any = [True] * (end + 1) __A : List[Any] = [] while start <= end: if temp[start] is True: in_prime.append(a ) for i in range(start * start , end + 1 , a ): __A : Optional[int] = False start += 1 prime += in_prime __A : Any = end + 1 __A : Any = min(2 * end , a ) while low <= n: __A : List[Any] = [True] * (high - low + 1) for each in in_prime: __A : List[str] = math.floor(low / each ) * each if t < low: t += each for j in range(a , high + 1 , a ): __A : Optional[int] = False for j in range(len(a ) ): if temp[j] is True: prime.append(j + low ) __A : Optional[int] = high + 1 __A : Tuple = min(high + end , a ) return prime print(sieve(10**6))
280
0
from __future__ import annotations import numpy as np def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = np.shape(lowerCAmelCase__ ) if rows != columns: lowercase = ( '''\'table\' has to be of square shaped array but got a ''' f'{rows}x{columns} array:\n{table}' ) raise ValueError(lowerCAmelCase__ ) lowercase = np.zeros((rows, columns) ) lowercase = np.zeros((rows, columns) ) for i in range(lowerCAmelCase__ ): for j in range(lowerCAmelCase__ ): lowercase = sum(lower[i][k] * upper[k][j] for k in range(lowerCAmelCase__ ) ) if upper[j][j] == 0: raise ArithmeticError('''No LU decomposition exists''' ) lowercase = (table[i][j] - total) / upper[j][j] lowercase = 1 for j in range(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = sum(lower[i][k] * upper[k][j] for k in range(lowerCAmelCase__ ) ) lowercase = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
101
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : str = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Union[str, Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} SCREAMING_SNAKE_CASE : int = { """vocab_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model""", } } SCREAMING_SNAKE_CASE : Union[str, Any] = { """camembert-base""": 512, } SCREAMING_SNAKE_CASE : List[Any] = """▁""" class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =['input_ids', 'attention_mask'] def __init__(self , a_ , a_="<s>" , a_="</s>" , a_="</s>" , a_="<s>" , a_="<unk>" , a_="<pad>" , a_="<mask>" , a_=["<s>NOTUSED", "</s>NOTUSED"] , a_ = None , **a_ , ): '''simple docstring''' __snake_case : List[Any] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else mask_token __snake_case : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , cls_token=a_ , pad_token=a_ , mask_token=a_ , additional_special_tokens=a_ , sp_model_kwargs=self.sp_model_kwargs , **a_ , ) __snake_case : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a_ ) ) __snake_case : Optional[Any] = vocab_file # HACK: These tokens were added by fairseq but don't seem to be actually used when duplicated in the actual # sentencepiece vocabulary (this is the case for <s> and </s> __snake_case : List[Any] = {'''<s>NOTUSED''': 0, '''<pad>''': 1, '''</s>NOTUSED''': 2, '''<unk>''': 3} __snake_case : str = len(self.fairseq_tokens_to_ids ) __snake_case : List[Any] = len(self.sp_model ) + len(self.fairseq_tokens_to_ids ) __snake_case : List[str] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __snake_case : Optional[Any] = [self.cls_token_id] __snake_case : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE (self , a_ , a_ = None , a_ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a_ , token_ids_a=a_ , already_has_special_tokens=a_ ) if token_ids_a is None: return [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1, 1] + ([0] * len(a_ )) + [1] def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' __snake_case : Optional[int] = [self.sep_token_id] __snake_case : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return len(self.fairseq_tokens_to_ids ) + len(self.sp_model ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Dict = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' return self.sp_model.encode(a_ , out_type=a_ ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] elif self.sp_model.PieceToId(a_ ) == 0: # Convert sentence piece unk token to fairseq unk token index return self.unk_token_id return self.fairseq_offset + self.sp_model.PieceToId(a_ ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' __snake_case : Tuple = [] __snake_case : Dict = '''''' __snake_case : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(a_ ) + token __snake_case : Tuple = True __snake_case : Any = [] else: current_sub_tokens.append(a_ ) __snake_case : int = False out_string += self.sp_model.decode(a_ ) return out_string.strip() def __getstate__(self ): '''simple docstring''' __snake_case : Dict = self.__dict__.copy() __snake_case : Union[str, Any] = None return state def __setstate__(self , a_ ): '''simple docstring''' __snake_case : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __snake_case : Optional[Any] = {} __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' if not os.path.isdir(a_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __snake_case : List[Any] = os.path.join( a_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a_ ) elif not os.path.isfile(self.vocab_file ): with open(a_ , '''wb''' ) as fi: __snake_case : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(a_ ) return (out_vocab_file,)
102
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __snake_case ( unittest.TestCase ): def __init__( self : int , A_ : List[str] , A_ : Dict=1_3 , A_ : Dict=7 , A_ : int=True , A_ : Optional[int]=True , A_ : List[str]=True , A_ : Union[str, Any]=True , A_ : Any=9_9 , A_ : List[str]=3_2 , A_ : int=5 , A_ : str=4 , A_ : Any=3_7 , A_ : Dict="gelu" , A_ : Any=0.1 , A_ : Any=0.1 , A_ : Tuple=5_1_2 , A_ : Optional[int]=1_6 , A_ : str=2 , A_ : Optional[int]=0.02 , A_ : Optional[int]=4 , ): lowerCAmelCase_ : List[Any] = parent lowerCAmelCase_ : Tuple = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : str = is_training lowerCAmelCase_ : Optional[Any] = use_attention_mask lowerCAmelCase_ : Optional[int] = use_token_type_ids lowerCAmelCase_ : str = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : str = num_attention_heads lowerCAmelCase_ : Any = intermediate_size lowerCAmelCase_ : Any = hidden_act lowerCAmelCase_ : Dict = hidden_dropout_prob lowerCAmelCase_ : Dict = attention_probs_dropout_prob lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : List[Any] = type_vocab_size lowerCAmelCase_ : List[str] = type_sequence_label_size lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Any = num_choices def UpperCAmelCase__ ( self : Tuple): lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowerCAmelCase_ : Dict = None if self.use_attention_mask: lowerCAmelCase_ : Optional[int] = random_attention_mask([self.batch_size, self.seq_length]) lowerCAmelCase_ : str = None if self.use_token_type_ids: lowerCAmelCase_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) lowerCAmelCase_ : Optional[int] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase__ ( self : Tuple): lowerCAmelCase_ : str = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Tuple = config_and_inputs lowerCAmelCase_ : List[str] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict def UpperCAmelCase__ ( self : Optional[int]): lowerCAmelCase_ : Any = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = config_and_inputs lowerCAmelCase_ : Tuple = True lowerCAmelCase_ : Union[str, Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) lowerCAmelCase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __snake_case ( UpperCamelCase_ ,unittest.TestCase ): _a = True _a = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self : List[str]): lowerCAmelCase_ : Optional[int] = FlaxRobertaPreLayerNormModelTester(self) @slow def UpperCAmelCase__ ( self : str): for model_class_name in self.all_model_classes: lowerCAmelCase_ : Tuple = model_class_name.from_pretrained('''andreasmadsen/efficient_mlm_m0.40''' , from_pt=A_) lowerCAmelCase_ : int = model(np.ones((1, 1))) self.assertIsNotNone(A_) @require_flax class __snake_case ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int): lowerCAmelCase_ : str = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained('''andreasmadsen/efficient_mlm_m0.40''' , from_pt=A_) lowerCAmelCase_ : Dict = np.array([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] , dtype=jnp.intaa) lowerCAmelCase_ : Union[str, Any] = model(A_)[0] lowerCAmelCase_ : Any = [1, 1_1, 5_0_2_6_5] self.assertEqual(list(output.shape) , A_) # compare the actual values for a slice. lowerCAmelCase_ : Union[str, Any] = np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa) self.assertTrue(np.allclose(output[:, :3, :3] , A_ , atol=1e-4)) @slow def UpperCAmelCase__ ( self : int): lowerCAmelCase_ : List[str] = FlaxRobertaPreLayerNormModel.from_pretrained('''andreasmadsen/efficient_mlm_m0.40''' , from_pt=A_) lowerCAmelCase_ : str = np.array([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] , dtype=jnp.intaa) lowerCAmelCase_ : List[Any] = model(A_)[0] # compare the actual values for a slice. lowerCAmelCase_ : Any = np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa) self.assertTrue(np.allclose(output[:, :3, :3] , A_ , atol=1e-4))
103
def _SCREAMING_SNAKE_CASE ( a , a = 0 ) -> list: __A : int = length or len(a ) __A : str = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __A , __A : Optional[int] = list_data[i + 1], list_data[i] __A : Union[str, Any] = True return list_data if not swapped else bubble_sort(a , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class lowercase_ : """simple docstring""" def __init__( self : str ,lowercase__ : int ,lowercase__ : List[str]=1_3 ,lowercase__ : List[str]=1_0 ,lowercase__ : int=3 ,lowercase__ : Tuple=2 ,lowercase__ : Union[str, Any]=2 ,lowercase__ : List[str]=2 ,lowercase__ : List[Any]=True ,lowercase__ : Any=True ,lowercase__ : Optional[int]=3_2 ,lowercase__ : List[str]=5 ,lowercase__ : Tuple=4 ,lowercase__ : str=3_7 ,lowercase__ : List[Any]="gelu" ,lowercase__ : Dict=0.1 ,lowercase__ : Any=0.1 ,lowercase__ : str=1_0 ,lowercase__ : Any=0.0_2 ,lowercase__ : Tuple=0.9 ,lowercase__ : Tuple=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = patch_size __lowercase = tubelet_size __lowercase = num_frames __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = mask_ratio __lowercase = scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame __lowercase = (image_size // patch_size) ** 2 __lowercase = (num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos __lowercase = int(mask_ratio * self.seq_length ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : List[str] ): return VideoMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,num_frames=self.num_frames ,tubelet_size=self.tubelet_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int ,lowercase__ : Any ,lowercase__ : Optional[Any] ): __lowercase = VideoMAEModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : Any ): __lowercase = VideoMAEForPreTraining(lowercase__ ) model.to(lowercase__ ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch __lowercase = torch.ones((self.num_masks,) ) __lowercase = torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) __lowercase = mask.expand(self.batch_size ,-1 ).bool() __lowercase = model(lowercase__ ,lowercase__ ) # model only returns predictions for masked patches __lowercase = mask.sum().item() __lowercase = 3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_masked_patches, decoder_num_labels) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : int = ( {'feature-extraction': VideoMAEModel, 'video-classification': VideoMAEForVideoClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : Any = False SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : Union[str, Any] = False def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = VideoMAEModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Tuple ,lowercase__ : List[str] ,lowercase__ : List[str]=False ): __lowercase = copy.deepcopy(lowercase__ ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch __lowercase = torch.ones((self.model_tester.num_masks,) ) __lowercase = torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) __lowercase = mask.expand(self.model_tester.batch_size ,-1 ).bool() __lowercase = bool_masked_pos.to(lowercase__ ) if return_labels: if model_class in [ *get_values(lowercase__ ), ]: __lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=lowercase__ ) return inputs_dict def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() @unittest.skip(reason='''VideoMAE does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Any ): for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = VideoMAEModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): if not self.has_attentions: pass else: __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: __lowercase = self.model_tester.seq_length - self.model_tester.num_masks __lowercase = ( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) __lowercase = True __lowercase = False __lowercase = True __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase__ ) ,self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowercase = True __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase__ ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) __lowercase = len(lowercase__ ) # Check attention is always last and order is fine __lowercase = True __lowercase = True __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) self.assertEqual(out_len + 1 ,len(lowercase__ ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase__ ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, seq_len, seq_len] ,) def SCREAMING_SNAKE_CASE ( self : Dict ): def check_hidden_states_output(lowercase__ : List[str] ,lowercase__ : int ,lowercase__ : int ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowercase__ ) ,lowercase__ ) __lowercase = self.model_tester.seq_length - self.model_tester.num_masks __lowercase = num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[seq_length, self.model_tester.hidden_size] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass def _A ( ): """simple docstring""" __lowercase = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) __lowercase = np.load(A__ ) return list(A__ ) @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : List[str] ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] ,image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = VideoMAEForVideoClassification.from_pretrained('''MCG-NJU/videomae-base-finetuned-kinetics''' ).to( lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_video() __lowercase = image_processor(lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 4_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_video() __lowercase = image_processor(lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # add boolean mask, indicating which patches to mask __lowercase = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' ,filename='''bool_masked_pos.pt''' ) __lowercase = torch.load(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor( [[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ,device=lowercase__ ) self.assertEqual(outputs.logits.shape ,lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] ,lowercase__ ,atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) __lowercase = torch.tensor([0.5_1_4_2] ,device=lowercase__ ) self.assertTrue(torch.allclose(outputs.loss ,lowercase__ ,atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) __lowercase = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ,norm_pix_loss=lowercase__ ).to( lowercase__ ) with torch.no_grad(): __lowercase = model(**lowercase__ ) __lowercase = torch.tensor(torch.tensor([0.6_4_6_9] ) ,device=lowercase__ ) self.assertTrue(torch.allclose(outputs.loss ,lowercase__ ,atol=1e-4 ) )
104
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a ) -> int: if not nums: return 0 __A : Optional[int] = nums[0] __A : str = 0 for num in nums[1:]: __A , __A : Tuple = ( max_excluding + num, max(a , a ), ) return max(a , a ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a : Tuple = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
105
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple ,lowercase_ : Dict ,lowercase_ : int=7 ,lowercase_ : str=3 ,lowercase_ : List[Any]=1_8 ,lowercase_ : Any=3_0 ,lowercase_ : str=4_0_0 ,lowercase_ : List[str]=True ,lowercase_ : int=None ,lowercase_ : Tuple=True ,lowercase_ : Union[str, Any]=False ,lowercase_ : Union[str, Any]=True ,lowercase_ : Union[str, Any]=True ,lowercase_ : str=[0.5, 0.5, 0.5] ,lowercase_ : Any=[0.5, 0.5, 0.5] ,): lowerCAmelCase__ : Any = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : Any = num_channels lowerCAmelCase__ : Tuple = image_size lowerCAmelCase__ : str = min_resolution lowerCAmelCase__ : Optional[Any] = max_resolution lowerCAmelCase__ : str = do_resize lowerCAmelCase__ : Dict = size if size is not None else {'''height''': 1_8, '''width''': 2_0} lowerCAmelCase__ : Union[str, Any] = do_thumbnail lowerCAmelCase__ : List[Any] = do_align_axis lowerCAmelCase__ : List[str] = do_pad lowerCAmelCase__ : List[str] = do_normalize lowerCAmelCase__ : Union[str, Any] = image_mean lowerCAmelCase__ : Tuple = image_std def __lowerCAmelCase ( self : List[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowercase__ = DonutImageProcessor if is_vision_available() else None def __lowerCAmelCase ( self : List[str] ): lowerCAmelCase__ : Dict = DonutImageProcessingTester(self ) @property def __lowerCAmelCase ( self : Optional[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : str ): lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase_ ,'''size''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_thumbnail''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_align_long_axis''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_pad''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_normalize''' ) ) self.assertTrue(hasattr(lowercase_ ,'''image_mean''' ) ) self.assertTrue(hasattr(lowercase_ ,'''image_std''' ) ) def __lowerCAmelCase ( self : Tuple ): lowerCAmelCase__ : Any = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''height''': 1_8, '''width''': 2_0} ) lowerCAmelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ) self.assertEqual(image_processor.size ,{'''height''': 4_2, '''width''': 4_2} ) # Previous config had dimensions in (width, height) order lowerCAmelCase__ : int = self.image_processing_class.from_dict(self.image_processor_dict ,size=(4_2, 8_4) ) self.assertEqual(image_processor.size ,{'''height''': 8_4, '''width''': 4_2} ) def __lowerCAmelCase ( self : Optional[int] ): pass @is_flaky() def __lowerCAmelCase ( self : Optional[Any] ): # Initialize image_processing lowerCAmelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : Tuple = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ ,Image.Image ) # Test not batched input lowerCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched lowerCAmelCase__ : List[Any] = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) @is_flaky() def __lowerCAmelCase ( self : Tuple ): # Initialize image_processing lowerCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ,numpify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ ,np.ndarray ) # Test not batched input lowerCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched lowerCAmelCase__ : Optional[Any] = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) @is_flaky() def __lowerCAmelCase ( self : str ): # Initialize image_processing lowerCAmelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ,torchify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ ,torch.Tensor ) # Test not batched input lowerCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched lowerCAmelCase__ : Any = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,)
106
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class snake_case__ (_UpperCamelCase ): """simple docstring""" def __init__( self : Optional[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=13 , __lowerCamelCase : List[str]=7 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Dict=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=99 , __lowerCamelCase : List[Any]=32 , __lowerCamelCase : Any=5 , __lowerCamelCase : str=4 , __lowerCamelCase : List[Any]=64 , __lowerCamelCase : Dict="gelu" , __lowerCamelCase : int=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Union[str, Any]=5_12 , __lowerCamelCase : List[Any]=16 , __lowerCamelCase : str=2 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Optional[int]=3 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=2 , __lowerCamelCase : List[Any]=2 , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : str=2 , __lowerCamelCase : Dict=4 , __lowerCamelCase : Optional[int]=1 , ) -> Any: a = parent a = batch_size a = seq_length a = is_training a = use_input_mask a = use_token_type_ids a = use_labels a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = type_sequence_label_size a = initializer_range a = num_labels a = num_choices a = scope a = q_groups a = k_groups a = v_groups a = post_attention_groups a = intermediate_groups a = output_groups def __UpperCAmelCase ( self : int ) -> Optional[int]: a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a = None if self.use_input_mask: a = random_attention_mask([self.batch_size, self.seq_length] ) a = None a = None a = None if self.use_labels: a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a = ids_tensor([self.batch_size] , self.num_choices ) a = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self : Dict ) -> Dict: return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def __UpperCAmelCase ( self : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> str: a = SqueezeBertModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = model(__lowerCamelCase , __lowerCamelCase ) a = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any , __lowerCamelCase : int , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] ) -> Dict: a = SqueezeBertForMaskedLM(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : str ) -> Union[str, Any]: a = SqueezeBertForQuestionAnswering(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = model( __lowerCamelCase , attention_mask=__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCAmelCase ( self : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : str ) -> Any: a = self.num_labels a = SqueezeBertForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple ) -> Any: a = self.num_labels a = SqueezeBertForTokenClassification(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple ) -> Optional[int]: a = self.num_choices a = SqueezeBertForMultipleChoice(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() a = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() a = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() a = model( __lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCAmelCase ( self : Any ) -> Optional[int]: a = self.prepare_config_and_inputs() ((a) , (a) , (a) , (a) , (a) , (a)) = config_and_inputs a = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case__ (_UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE_ : Tuple = ( { """feature-extraction""": SqueezeBertModel, """fill-mask""": SqueezeBertForMaskedLM, """question-answering""": SqueezeBertForQuestionAnswering, """text-classification""": SqueezeBertForSequenceClassification, """token-classification""": SqueezeBertForTokenClassification, """zero-shot""": SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = True SCREAMING_SNAKE_CASE_ : Optional[int] = False def __UpperCAmelCase ( self : Tuple ) -> Any: a = SqueezeBertModelTester(self ) a = ConfigTester(self , config_class=__lowerCamelCase , dim=37 ) def __UpperCAmelCase ( self : int ) -> int: self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Tuple ) -> Dict: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[int] ) -> Dict: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[Any] ) -> Optional[Any]: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*__lowerCamelCase ) def __UpperCAmelCase ( self : Dict ) -> Any: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*__lowerCamelCase ) def __UpperCAmelCase ( self : Union[str, Any] ) -> str: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*__lowerCamelCase ) @slow def __UpperCAmelCase ( self : Any ) -> Any: for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a = SqueezeBertModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @require_sentencepiece @require_tokenizers @require_torch class snake_case__ (unittest.TestCase ): """simple docstring""" @slow def __UpperCAmelCase ( self : str ) -> Union[str, Any]: a = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli" ) a = torch.tensor([[1, 2_94_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 13, 15_88, 2]] ) a = model(__lowerCamelCase )[0] a = torch.Size((1, 3) ) self.assertEqual(output.shape , __lowerCamelCase ) a = torch.tensor([[0.6_401, -0.0_349, -0.6_041]] ) self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) )
107
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> None: __A : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a ) == len(a ), F"""{len(a )} != {len(a )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) UpperCAmelCase : List[Any] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } UpperCAmelCase : Optional[int] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _SCREAMING_SNAKE_CASE ( a , a ) -> Dict: try: __A : int = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a ) ) def _SCREAMING_SNAKE_CASE ( a , a ) -> List[int]: if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _SCREAMING_SNAKE_CASE ( a , a = "student" , a = None , a = None , a=False , a=None , a=None , **a , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __A : List[str] = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(a , a ): AutoTokenizer.from_pretrained(a ).save_pretrained(a ) # purely for convenience __A : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained(a ).eval() else: assert isinstance(a , a ), F"""teacher must be a model or string got type {type(a )}""" __A : int = teacher.config.to_diff_dict() try: __A , __A : List[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A : str = teacher_e if d is None: __A : List[Any] = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): __A , __A : List[Any] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A : Optional[int] = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A : int = teacher_e if d is None: __A : Optional[Any] = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a ) # Copy weights __A : Dict = teacher.config_class(**a ) __A : int = AutoModelForSeqaSeqLM.from_config(a ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A : Any = student.load_state_dict(teacher.state_dict() , strict=a ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A : Optional[int] = list(range(a ) ), list(range(a ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) if d_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) try: if hasattr( a , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a ) copy_layers(teacher.decoder.block , student.decoder.block , a ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __A : Optional[int] = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(a ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
280
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
108
def _SCREAMING_SNAKE_CASE ( a , a ) -> list[int]: __A : Optional[int] = int(a ) # Initialize Result __A : Optional[int] = [] # Traverse through all denomination for denomination in reversed(a ): # Find denominations while int(a ) >= int(a ): total_value -= int(a ) answer.append(a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase : List[str] = [] UpperCAmelCase : Optional[int] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): UpperCAmelCase : List[Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) UpperCAmelCase : int = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCAmelCase : Tuple = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(F"""Following is minimal change for {value}: """) UpperCAmelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
280
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging A: str = logging.get_logger(__name__) A: str = "▁" A: Union[str, Any] = {"vocab_file": "sentencepiece.bpe.model"} A: Union[str, Any] = { "vocab_file": { "facebook/nllb-200-distilled-600M": ( "https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model" ), } } A: Dict = { "facebook/nllb-200-distilled-600M": 1_0_2_4, } # fmt: off A: List[Any] = ["ace_Arab", "ace_Latn", "acm_Arab", "acq_Arab", "aeb_Arab", "afr_Latn", "ajp_Arab", "aka_Latn", "amh_Ethi", "apc_Arab", "arb_Arab", "ars_Arab", "ary_Arab", "arz_Arab", "asm_Beng", "ast_Latn", "awa_Deva", "ayr_Latn", "azb_Arab", "azj_Latn", "bak_Cyrl", "bam_Latn", "ban_Latn", "bel_Cyrl", "bem_Latn", "ben_Beng", "bho_Deva", "bjn_Arab", "bjn_Latn", "bod_Tibt", "bos_Latn", "bug_Latn", "bul_Cyrl", "cat_Latn", "ceb_Latn", "ces_Latn", "cjk_Latn", "ckb_Arab", "crh_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "dik_Latn", "dyu_Latn", "dzo_Tibt", "ell_Grek", "eng_Latn", "epo_Latn", "est_Latn", "eus_Latn", "ewe_Latn", "fao_Latn", "pes_Arab", "fij_Latn", "fin_Latn", "fon_Latn", "fra_Latn", "fur_Latn", "fuv_Latn", "gla_Latn", "gle_Latn", "glg_Latn", "grn_Latn", "guj_Gujr", "hat_Latn", "hau_Latn", "heb_Hebr", "hin_Deva", "hne_Deva", "hrv_Latn", "hun_Latn", "hye_Armn", "ibo_Latn", "ilo_Latn", "ind_Latn", "isl_Latn", "ita_Latn", "jav_Latn", "jpn_Jpan", "kab_Latn", "kac_Latn", "kam_Latn", "kan_Knda", "kas_Arab", "kas_Deva", "kat_Geor", "knc_Arab", "knc_Latn", "kaz_Cyrl", "kbp_Latn", "kea_Latn", "khm_Khmr", "kik_Latn", "kin_Latn", "kir_Cyrl", "kmb_Latn", "kon_Latn", "kor_Hang", "kmr_Latn", "lao_Laoo", "lvs_Latn", "lij_Latn", "lim_Latn", "lin_Latn", "lit_Latn", "lmo_Latn", "ltg_Latn", "ltz_Latn", "lua_Latn", "lug_Latn", "luo_Latn", "lus_Latn", "mag_Deva", "mai_Deva", "mal_Mlym", "mar_Deva", "min_Latn", "mkd_Cyrl", "plt_Latn", "mlt_Latn", "mni_Beng", "khk_Cyrl", "mos_Latn", "mri_Latn", "zsm_Latn", "mya_Mymr", "nld_Latn", "nno_Latn", "nob_Latn", "npi_Deva", "nso_Latn", "nus_Latn", "nya_Latn", "oci_Latn", "gaz_Latn", "ory_Orya", "pag_Latn", "pan_Guru", "pap_Latn", "pol_Latn", "por_Latn", "prs_Arab", "pbt_Arab", "quy_Latn", "ron_Latn", "run_Latn", "rus_Cyrl", "sag_Latn", "san_Deva", "sat_Beng", "scn_Latn", "shn_Mymr", "sin_Sinh", "slk_Latn", "slv_Latn", "smo_Latn", "sna_Latn", "snd_Arab", "som_Latn", "sot_Latn", "spa_Latn", "als_Latn", "srd_Latn", "srp_Cyrl", "ssw_Latn", "sun_Latn", "swe_Latn", "swh_Latn", "szl_Latn", "tam_Taml", "tat_Cyrl", "tel_Telu", "tgk_Cyrl", "tgl_Latn", "tha_Thai", "tir_Ethi", "taq_Latn", "taq_Tfng", "tpi_Latn", "tsn_Latn", "tso_Latn", "tuk_Latn", "tum_Latn", "tur_Latn", "twi_Latn", "tzm_Tfng", "uig_Arab", "ukr_Cyrl", "umb_Latn", "urd_Arab", "uzn_Latn", "vec_Latn", "vie_Latn", "war_Latn", "wol_Latn", "xho_Latn", "ydd_Hebr", "yor_Latn", "yue_Hant", "zho_Hans", "zho_Hant", "zul_Latn"] class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : str = VOCAB_FILES_NAMES __lowerCAmelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase : int = ['input_ids', 'attention_mask'] __lowerCAmelCase : List[int] = [] __lowerCAmelCase : List[int] = [] def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE="<mask>" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False , **_SCREAMING_SNAKE_CASE , ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Tuple = AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else mask_token UpperCAmelCase : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase : Any = legacy_behaviour super().__init__( bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , src_lang=_SCREAMING_SNAKE_CASE , tgt_lang=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , legacy_behaviour=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_SCREAMING_SNAKE_CASE ) ) UpperCAmelCase : List[str] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase : Dict = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase : str = 1 UpperCAmelCase : Optional[Any] = len(self.sp_model ) UpperCAmelCase : Dict = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(_SCREAMING_SNAKE_CASE ) } UpperCAmelCase : Any = {v: k for k, v in self.lang_code_to_id.items()} UpperCAmelCase : Any = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCAmelCase : Any = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCAmelCase : Optional[int] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCAmelCase : List[str] = src_lang if src_lang is not None else """eng_Latn""" UpperCAmelCase : str = self.lang_code_to_id[self._src_lang] UpperCAmelCase : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : int = self.__dict__.copy() UpperCAmelCase : Optional[Any] = None UpperCAmelCase : Tuple = self.sp_model.serialized_model_proto() return state def __setstate__( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCAmelCase : Union[str, Any] = {} UpperCAmelCase : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> None: '''simple docstring''' UpperCAmelCase : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_SCREAMING_SNAKE_CASE , token_ids_a=_SCREAMING_SNAKE_CASE , already_has_special_tokens=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[str] = [1] * len(self.prefix_tokens ) UpperCAmelCase : str = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(_SCREAMING_SNAKE_CASE )) + suffix_ones return prefix_ones + ([0] * len(_SCREAMING_SNAKE_CASE )) + ([0] * len(_SCREAMING_SNAKE_CASE )) + suffix_ones def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[int]: '''simple docstring''' UpperCAmelCase : Union[str, Any] = [self.sep_token_id] UpperCAmelCase : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) UpperCAmelCase : Any = src_lang UpperCAmelCase : Dict = self(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) UpperCAmelCase : str = self.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = tgt_lang_id return inputs def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : Any = {self.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' return self.sp_model.encode(_SCREAMING_SNAKE_CASE , out_type=_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[int] = self.sp_model.PieceToId(_SCREAMING_SNAKE_CASE ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : int = """""".join(_SCREAMING_SNAKE_CASE ).replace(_SCREAMING_SNAKE_CASE , """ """ ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase : int = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(_SCREAMING_SNAKE_CASE , """wb""" ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(_SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = "eng_Latn" , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = "fra_Latn" , **_SCREAMING_SNAKE_CASE , ) -> BatchEncoding: '''simple docstring''' UpperCAmelCase : List[str] = src_lang UpperCAmelCase : Tuple = tgt_lang return super().prepare_seqaseq_batch(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> None: '''simple docstring''' UpperCAmelCase : Optional[Any] = self.lang_code_to_id[src_lang] if self.legacy_behaviour: UpperCAmelCase : Optional[int] = [] UpperCAmelCase : List[Any] = [self.eos_token_id, self.cur_lang_code] else: UpperCAmelCase : List[Any] = [self.cur_lang_code] UpperCAmelCase : Optional[Any] = [self.eos_token_id] def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ) -> None: '''simple docstring''' UpperCAmelCase : List[str] = self.lang_code_to_id[lang] if self.legacy_behaviour: UpperCAmelCase : Any = [] UpperCAmelCase : List[Any] = [self.eos_token_id, self.cur_lang_code] else: UpperCAmelCase : Tuple = [self.cur_lang_code] UpperCAmelCase : Optional[int] = [self.eos_token_id]
109
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _A( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __A : List[Any] = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __A : Union[str, Any] = parent __A : Optional[int] = batch_size __A : int = num_channels __A : int = min_resolution __A : Any = max_resolution __A : List[Any] = do_resize __A : List[Any] = size __A : Union[str, Any] = do_normalize __A : Optional[int] = image_mean __A : Optional[int] = image_std __A : int = do_rescale __A : str = rescale_factor __A : Tuple = do_pad def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase_ ( self , _A , _A=False ): if not batched: __A : List[str] = image_inputs[0] if isinstance(_A , Image.Image ): __A , __A : int = image.size else: __A , __A : Any = image.shape[1], image.shape[2] if w < h: __A : List[Any] = int(self.size['shortest_edge'] * h / w ) __A : List[Any] = self.size['shortest_edge'] elif w > h: __A : Union[str, Any] = self.size['shortest_edge'] __A : str = int(self.size['shortest_edge'] * w / h ) else: __A : Dict = self.size['shortest_edge'] __A : str = self.size['shortest_edge'] else: __A : int = [] for image in image_inputs: __A , __A : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __A : List[str] = max(_A , key=lambda _A : item[0] )[0] __A : str = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : List[str] = YolosImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): __A : Dict = YolosImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): __A : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) def UpperCAmelCase_ ( self ): __A : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , _A ) __A : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _A ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing __A : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input __A : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A , __A : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) __A : str = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input __A : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : List[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Tuple = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input __A : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Union[str, Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Optional[int] = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processings __A : Tuple = self.image_processing_class(**self.image_processor_dict ) __A : Any = self.image_processing_class(do_resize=_A , do_normalize=_A , do_rescale=_A ) # create random PyTorch tensors __A : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __A : Optional[int] = image_processing_a.pad(_A , return_tensors='pt' ) __A : Optional[int] = image_processing_a(_A , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1e-4 ) ) @slow def UpperCAmelCase_ ( self ): # prepare image and target __A : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __A : Optional[Any] = json.loads(f.read() ) __A : Optional[Any] = {'image_id': 39769, 'annotations': target} # encode them __A : str = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) __A : List[Any] = image_processing(images=_A , annotations=_A , return_tensors='pt' ) # verify pixel values __A : List[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : List[Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Any = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify orig_size __A : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : str = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) ) @slow def UpperCAmelCase_ ( self ): # prepare image, target and masks_path __A : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __A : Tuple = json.loads(f.read() ) __A : Any = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __A : List[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __A : Any = YolosImageProcessor(format='coco_panoptic' ) __A : List[Any] = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='pt' ) # verify pixel values __A : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Union[str, Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : List[str] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify masks __A : Tuple = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _A ) # verify orig_size __A : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : int = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) )
280
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available lowercase__ = { '''configuration_longt5''': ['''LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongT5Config''', '''LongT5OnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ '''LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongT5EncoderModel''', '''LongT5ForConditionalGeneration''', '''LongT5Model''', '''LongT5PreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ '''FlaxLongT5ForConditionalGeneration''', '''FlaxLongT5Model''', '''FlaxLongT5PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
import argparse import json from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( ) -> List[Any]: __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=a , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=a , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=a , help='where to store parsed gold_data_path file' , ) __A : Optional[int] = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: __A : List[Any] = json.load(a ) for dpr_record in tqdm(a ): __A : Dict = dpr_record['question'] __A : Any = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(a ) + '\n' ) if __name__ == "__main__": main()
280
0
from PIL import Image def A ( _SCREAMING_SNAKE_CASE ) -> Image: lowerCamelCase : Union[str, Any] = image.size lowerCamelCase : Dict = 0 lowerCamelCase : Tuple = image.load() for i in range(_SCREAMING_SNAKE_CASE ): for j in range(_SCREAMING_SNAKE_CASE ): lowerCamelCase : Dict = pixels[j, i] mean += pixel mean //= width * height for j in range(_SCREAMING_SNAKE_CASE ): for i in range(_SCREAMING_SNAKE_CASE ): lowerCamelCase : Any = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Optional[Any] = mean_threshold(Image.open('path_to_image').convert('L')) image.save('output_image_path')
48
from heapq import heappop, heappush import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a , ) -> tuple[float | int, list[tuple[int, int]]]: __A , __A : int = grid.shape __A : Any = [-1, 1, 0, 0] __A : Optional[Any] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __A , __A : Optional[int] = [(0, source)], set() __A : Any = np.full((rows, cols) , np.inf ) __A : Any = 0 __A : Any = np.empty((rows, cols) , dtype=a ) __A : Optional[Any] = None while queue: ((__A) , (__A)) : List[str] = heappop(a ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __A : int = [] while (x, y) != source: path.append((x, y) ) __A , __A : Optional[int] = predecessors[x, y] path.append(a ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a ) ): __A , __A : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __A : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a , (dist + 1, (nx, ny)) ) __A : List[Any] = dist + 1 __A : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
280
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process lowercase = logging.getLogger(__name__) def __UpperCAmelCase ( a_ , a_): return (preds == labels).mean() @dataclass class UpperCamelCase_ : '''simple docstring''' lowerCAmelCase = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowerCAmelCase = field( default=snake_case__ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) lowerCAmelCase = field( default=snake_case__ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) lowerCAmelCase = field( default=snake_case__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class UpperCamelCase_ : '''simple docstring''' lowerCAmelCase = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(processors.keys() )} ) lowerCAmelCase = field(metadata={'''help''': '''Should contain the data files for the task.'''} ) lowerCAmelCase = field( default=1_2_8 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCAmelCase = field( default=snake_case__ , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def __UpperCAmelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) snake_case_ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir) and os.listdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' ' --overwrite_output_dir to overcome.') # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , a_) # Set seed set_seed(training_args.seed) try: snake_case_ = processors[data_args.task_name]() snake_case_ = processor.get_labels() snake_case_ = len(a_) except KeyError: raise ValueError('Task not found: %s' % (data_args.task_name)) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=a_ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path) , config=a_ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case_ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=a_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case_ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=a_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(a_) -> Dict: snake_case_ = np.argmax(p.predictions , axis=1) return {"acc": simple_accuracy(a_ , p.label_ids)} # Data collator snake_case_ = DataCollatorWithPadding(a_ , pad_to_multiple_of=8) if training_args.fpaa else None # Initialize our Trainer snake_case_ = Trainer( model=a_ , args=a_ , train_dataset=a_ , eval_dataset=a_ , compute_metrics=a_ , data_collator=a_ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path) else None) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir) # Evaluation snake_case_ = {} if training_args.do_eval: logger.info('*** Evaluate ***') snake_case_ = trainer.evaluate() snake_case_ = os.path.join(training_args.output_dir , 'eval_results.txt') if trainer.is_world_master(): with open(a_ , 'w') as writer: logger.info('***** Eval results *****') for key, value in result.items(): logger.info(' %s = %s' , a_ , a_) writer.write('%s = %s\n' % (key, value)) results.update(a_) return results def __UpperCAmelCase ( a_): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
178
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
0
'''simple docstring''' import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters UpperCamelCase_ = logging.get_logger(__name__) def _UpperCAmelCase ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : Tuple , _lowerCamelCase : List[Any]=None , _lowerCamelCase : Optional[int]=None ) -> Optional[Any]: # Recurse if needed if "." in tensor_name: _lowerCAmelCase : List[Any] = tensor_name.split(""".""" ) for split in splits[:-1]: _lowerCAmelCase : int = getattr(_lowerCamelCase , _lowerCamelCase ) if new_module is None: raise ValueError(f'{module} has no attribute {split}.' ) _lowerCAmelCase : Tuple = new_module _lowerCAmelCase : int = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f'{module} does not have a parameter or a buffer named {tensor_name}.' ) _lowerCAmelCase : Tuple = tensor_name in module._buffers _lowerCAmelCase : Tuple = getattr(_lowerCamelCase , _lowerCamelCase ) if old_value.device == torch.device("""meta""" ) and device not in ["meta", torch.device("""meta""" )] and value is None: raise ValueError(f'{tensor_name} is on the meta device, we need a `value` to put in on {device}.' ) _lowerCAmelCase : Optional[Any] = False _lowerCAmelCase : List[str] = False if is_buffer or not is_bitsandbytes_available(): _lowerCAmelCase : List[Any] = False _lowerCAmelCase : Dict = False else: _lowerCAmelCase : Optional[Any] = hasattr(bnb.nn , """Params4bit""" ) and isinstance(module._parameters[tensor_name] , bnb.nn.Paramsabit ) _lowerCAmelCase : Optional[int] = isinstance(module._parameters[tensor_name] , bnb.nn.IntaParams ) if is_abit or is_abit: _lowerCAmelCase : Optional[int] = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: _lowerCAmelCase : int = old_value.to(_lowerCamelCase ) elif isinstance(_lowerCamelCase , torch.Tensor ): _lowerCAmelCase : Dict = value.to("""cpu""" ) if value.dtype == torch.inta: _lowerCAmelCase : List[str] = version.parse(importlib.metadata.version("""bitsandbytes""" ) ) > version.parse( """0.37.2""" ) if not is_abit_serializable: raise ValueError( """Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. """ """Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.""" ) else: _lowerCAmelCase : Union[str, Any] = torch.tensor(_lowerCamelCase , device="""cpu""" ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls , _lowerCamelCase ) and fpaa_statistics is None: _lowerCAmelCase : Any = new_value.T _lowerCAmelCase : Union[str, Any] = old_value.__dict__ if is_abit: _lowerCAmelCase : Optional[Any] = bnb.nn.IntaParams(_lowerCamelCase , requires_grad=_lowerCamelCase , **_lowerCamelCase ).to(_lowerCamelCase ) elif is_abit: _lowerCAmelCase : Tuple = bnb.nn.Paramsabit(_lowerCamelCase , requires_grad=_lowerCamelCase , **_lowerCamelCase ).to(_lowerCamelCase ) _lowerCAmelCase : Optional[int] = new_value if fpaa_statistics is not None: setattr(module.weight , """SCB""" , fpaa_statistics.to(_lowerCamelCase ) ) else: if value is None: _lowerCAmelCase : Dict = old_value.to(_lowerCamelCase ) elif isinstance(_lowerCamelCase , torch.Tensor ): _lowerCAmelCase : Union[str, Any] = value.to(_lowerCamelCase ) else: _lowerCAmelCase : str = torch.tensor(_lowerCamelCase , device=_lowerCamelCase ) if is_buffer: _lowerCAmelCase : List[Any] = new_value else: _lowerCAmelCase : str = nn.Parameter(_lowerCamelCase , requires_grad=old_value.requires_grad ) _lowerCAmelCase : Optional[int] = new_value def _UpperCAmelCase ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : int=None , _lowerCamelCase : Any=None , _lowerCamelCase : Optional[int]=None , _lowerCamelCase : Optional[Any]=False ) -> str: for name, module in model.named_children(): if current_key_name is None: _lowerCAmelCase : Optional[Any] = [] current_key_name.append(_lowerCamelCase ) if (isinstance(_lowerCamelCase , nn.Linear ) or isinstance(_lowerCamelCase , _lowerCamelCase )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in """.""".join(_lowerCamelCase ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(_lowerCamelCase , _lowerCamelCase ): _lowerCAmelCase : str = module.weight.shape else: _lowerCAmelCase : int = module.in_features _lowerCAmelCase : List[Any] = module.out_features if quantization_config.quantization_method() == "llm_int8": _lowerCAmelCase : int = bnb.nn.LinearabitLt( _lowerCamelCase , _lowerCamelCase , module.bias is not None , has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight , threshold=quantization_config.llm_inta_threshold , ) _lowerCAmelCase : List[Any] = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: _lowerCAmelCase : Any = bnb.nn.Linearabit( _lowerCamelCase , _lowerCamelCase , module.bias is not None , quantization_config.bnb_abit_compute_dtype , compress_statistics=quantization_config.bnb_abit_use_double_quant , quant_type=quantization_config.bnb_abit_quant_type , ) _lowerCAmelCase : Union[str, Any] = True # Store the module class in case we need to transpose the weight later _lowerCAmelCase : str = type(_lowerCamelCase ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(_lowerCamelCase ) if len(list(module.children() ) ) > 0: _lowerCAmelCase : str = _replace_with_bnb_linear( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , has_been_replaced=_lowerCamelCase , ) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def _UpperCAmelCase ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Optional[Any]=None , _lowerCamelCase : int=None , _lowerCamelCase : Tuple=None ) -> Any: _lowerCAmelCase : Optional[int] = ['lm_head'] if modules_to_not_convert is None else modules_to_not_convert _lowerCAmelCase : Optional[Any] = _replace_with_bnb_linear( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if not has_been_replaced: logger.warning( """You are loading your model in 8bit or 4bit but no linear modules were found in your model.""" """ Please double check your model architecture, or submit an issue on github if you think this is""" """ a bug.""" ) return model def _UpperCAmelCase ( *_lowerCamelCase : Any , **_lowerCamelCase : List[Any] ) -> str: warnings.warn( """`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead""" , _lowerCamelCase , ) return replace_with_bnb_linear(*_lowerCamelCase , **_lowerCamelCase ) def _UpperCAmelCase ( *_lowerCamelCase : str , **_lowerCamelCase : Optional[Any] ) -> Tuple: warnings.warn( """`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead""" , _lowerCamelCase , ) return set_module_quantized_tensor_to_device(*_lowerCamelCase , **_lowerCamelCase ) def _UpperCAmelCase ( _lowerCamelCase : str ) -> Dict: _lowerCAmelCase : str = deepcopy(_lowerCamelCase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() _lowerCAmelCase : Tuple = find_tied_parameters(_lowerCamelCase ) # For compatibility with Accelerate < 0.18 if isinstance(_lowerCamelCase , _lowerCamelCase ): _lowerCAmelCase : Optional[int] = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: _lowerCAmelCase : Dict = sum(_lowerCamelCase , [] ) _lowerCAmelCase : Optional[int] = len(_lowerCamelCase ) > 0 # Check if it is a base model _lowerCAmelCase : Dict = not hasattr(_lowerCamelCase , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head _lowerCAmelCase : Tuple = list(model.named_children() ) _lowerCAmelCase : Optional[int] = [list_modules[-1][0]] # add last module together with tied weights _lowerCAmelCase : List[Any] = set(_lowerCamelCase ) - set(_lowerCamelCase ) _lowerCAmelCase : int = list(set(_lowerCamelCase ) ) + list(_lowerCamelCase ) # remove ".weight" from the keys _lowerCAmelCase : List[Any] = ['.weight', '.bias'] _lowerCAmelCase : int = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: _lowerCAmelCase : int = name.replace(_lowerCamelCase , """""" ) filtered_module_names.append(_lowerCamelCase ) return filtered_module_names
309
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _SCREAMING_SNAKE_CASE ( a , a ) -> Optional[Any]: __A : Any = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 10_24, 'hidden_size': 7_68, 'max_length': 5_12, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 10_24, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } __A : str = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __A : Optional[int] = BERTEncoder( attention_cell=predefined_args['attention_cell'] , num_layers=predefined_args['num_layers'] , units=predefined_args['units'] , hidden_size=predefined_args['hidden_size'] , max_length=predefined_args['max_length'] , num_heads=predefined_args['num_heads'] , scaled=predefined_args['scaled'] , dropout=predefined_args['dropout'] , output_attention=a , output_all_encodings=a , use_residual=predefined_args['use_residual'] , activation=predefined_args.get('activation' , 'gelu' ) , layer_norm_eps=predefined_args.get('layer_norm_eps' , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __A : Union[str, Any] = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab __A : Any = os.path.join(get_home_dir() , 'models' ) __A : List[Any] = _load_vocab(a , a , a , cls=a ) __A : Dict = nlp.model.BERTModel( a , len(a ) , units=predefined_args['units'] , embed_size=predefined_args['embed_size'] , embed_dropout=predefined_args['embed_dropout'] , word_embed=predefined_args['word_embed'] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args['token_type_vocab_size'] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) __A : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __A : Any = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(a ), } __A : int = BertConfig.from_dict(a ) __A : Union[str, Any] = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a , a ): __A : Tuple = hf_param.shape __A : str = to_torch(params[gluon_param] ) __A : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), F"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __A : str = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , 'word_embed.0.weight' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , 'encoder.position_weight' ) __A : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , 'encoder.layer_norm.beta' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , 'encoder.layer_norm.gamma' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __A : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __A : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __A : BertSelfAttention = layer.attention.self __A : Optional[Any] = check_and_map_params( self_attn.key.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.key.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.query.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __A : Optional[Any] = check_and_map_params( self_attn.query.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.value.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.value.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __A : BertSelfOutput = layer.attention.output __A : Tuple = check_and_map_params( self_output.dense.bias , F"""encoder.transformer_cells.{i}.proj.bias""" ) __A : int = check_and_map_params( self_output.dense.weight , F"""encoder.transformer_cells.{i}.proj.weight""" ) __A : List[Any] = check_and_map_params( self_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __A : str = check_and_map_params( self_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __A : BertIntermediate = layer.intermediate __A : int = check_and_map_params( intermediate.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __A : List[Any] = check_and_map_params( intermediate.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __A : BertOutput = layer.output __A : List[Any] = check_and_map_params( bert_output.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __A : Dict = check_and_map_params( bert_output.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __A : Optional[int] = check_and_map_params( bert_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __A : Dict = check_and_map_params( bert_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __A : Any = RobertaTokenizer.from_pretrained('roberta-base' ) __A : List[str] = tokenizer.encode_plus(a )['input_ids'] # Get gluon output __A : List[str] = mx.nd.array([input_ids] ) __A : Union[str, Any] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) __A : Optional[Any] = BertModel.from_pretrained(a ) hf_bort_model.eval() __A : Tuple = tokenizer.encode_plus(a , return_tensors='pt' ) __A : Any = hf_bort_model(**a )[0] __A : Union[str, Any] = output_gluon[0].asnumpy() __A : Tuple = output_hf[0].detach().numpy() __A : int = np.max(np.abs(hf_layer - gluon_layer ) ).item() __A : int = np.allclose(a , a , atol=1e-3 ) if success: print('✔️ Both model do output the same tensors' ) else: print('❌ Both model do **NOT** output the same tensors' ) print('Absolute difference is:' , a ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase : Dict = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
280
0
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __lowerCAmelCase : def __init__( self , lowerCAmelCase , lowerCAmelCase=2 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=2 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=99 , lowerCAmelCase=36 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=16 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=6 , lowerCAmelCase=6 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=None , lowerCAmelCase=1_000 , ) -> Tuple: '''simple docstring''' _lowercase =parent _lowercase =batch_size _lowercase =num_channels _lowercase =image_size _lowercase =patch_size _lowercase =text_seq_length _lowercase =is_training _lowercase =use_input_mask _lowercase =use_token_type_ids _lowercase =use_labels _lowercase =vocab_size _lowercase =hidden_size _lowercase =num_hidden_layers _lowercase =num_attention_heads _lowercase =intermediate_size _lowercase =hidden_act _lowercase =hidden_dropout_prob _lowercase =attention_probs_dropout_prob _lowercase =max_position_embeddings _lowercase =type_vocab_size _lowercase =type_sequence_label_size _lowercase =initializer_range _lowercase =coordinate_size _lowercase =shape_size _lowercase =num_labels _lowercase =num_choices _lowercase =scope _lowercase =range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _lowercase =text_seq_length _lowercase =(image_size // patch_size) ** 2 + 1 _lowercase =self.text_seq_length + self.image_seq_length def A__ ( self ) -> int: '''simple docstring''' _lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _lowercase =ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _lowercase =bbox[i, j, 3] _lowercase =bbox[i, j, 1] _lowercase =t if bbox[i, j, 2] < bbox[i, j, 0]: _lowercase =bbox[i, j, 2] _lowercase =bbox[i, j, 0] _lowercase =t _lowercase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase =None if self.use_input_mask: _lowercase =random_attention_mask([self.batch_size, self.text_seq_length] ) _lowercase =None if self.use_token_type_ids: _lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _lowercase =None _lowercase =None if self.use_labels: _lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowercase =ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _lowercase =LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> str: '''simple docstring''' _lowercase =LayoutLMvaModel(config=_A ) model.to(_A ) model.eval() # text + image _lowercase =model(_A , pixel_values=_A ) _lowercase =model( _A , bbox=_A , pixel_values=_A , attention_mask=_A , token_type_ids=_A ) _lowercase =model(_A , bbox=_A , pixel_values=_A , token_type_ids=_A ) _lowercase =model(_A , bbox=_A , pixel_values=_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _lowercase =model(_A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _lowercase =model(pixel_values=_A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Tuple: '''simple docstring''' _lowercase =self.num_labels _lowercase =LayoutLMvaForSequenceClassification(_A ) model.to(_A ) model.eval() _lowercase =model( _A , bbox=_A , pixel_values=_A , attention_mask=_A , token_type_ids=_A , labels=_A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> List[str]: '''simple docstring''' _lowercase =self.num_labels _lowercase =LayoutLMvaForTokenClassification(config=_A ) model.to(_A ) model.eval() _lowercase =model( _A , bbox=_A , pixel_values=_A , attention_mask=_A , token_type_ids=_A , labels=_A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> List[Any]: '''simple docstring''' _lowercase =LayoutLMvaForQuestionAnswering(config=_A ) model.to(_A ) model.eval() _lowercase =model( _A , bbox=_A , pixel_values=_A , attention_mask=_A , token_type_ids=_A , start_positions=_A , end_positions=_A , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A__ ( self ) -> Tuple: '''simple docstring''' _lowercase =self.prepare_config_and_inputs() ( _lowercase ) =config_and_inputs _lowercase ={ 'input_ids': input_ids, 'bbox': bbox, 'pixel_values': pixel_values, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class __lowerCAmelCase ( snake_case__ , snake_case__ , unittest.TestCase ): _a = False _a = False _a = False _a = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) _a = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Any: '''simple docstring''' return True def A__ ( self ) -> Optional[Any]: '''simple docstring''' _lowercase =LayoutLMvaModelTester(self ) _lowercase =ConfigTester(self , config_class=_A , hidden_size=37 ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ) -> Union[str, Any]: '''simple docstring''' _lowercase =copy.deepcopy(_A ) if model_class in get_values(_A ): _lowercase ={ k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(_A , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(_A ): _lowercase =torch.ones(self.model_tester.batch_size , dtype=torch.long , device=_A ) elif model_class in get_values(_A ): _lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_A ) _lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_A ) elif model_class in [ *get_values(_A ), ]: _lowercase =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_A ) elif model_class in [ *get_values(_A ), ]: _lowercase =torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=_A , ) return inputs_dict def A__ ( self ) -> str: '''simple docstring''' self.config_tester.run_common_tests() def A__ ( self ) -> Dict: '''simple docstring''' _lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def A__ ( self ) -> Dict: '''simple docstring''' _lowercase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowercase =type self.model_tester.create_and_check_model(*_A ) def A__ ( self ) -> List[Any]: '''simple docstring''' _lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_A ) def A__ ( self ) -> Optional[int]: '''simple docstring''' _lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_A ) def A__ ( self ) -> Tuple: '''simple docstring''' _lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_A ) @slow def A__ ( self ) -> Dict: '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase =LayoutLMvaModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def a ( ) -> Any: """simple docstring""" _lowercase =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase ): @cached_property def A__ ( self ) -> Dict: '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=_A ) if is_vision_available() else None @slow def A__ ( self ) -> Union[str, Any]: '''simple docstring''' _lowercase =LayoutLMvaModel.from_pretrained('microsoft/layoutlmv3-base' ).to(_A ) _lowercase =self.default_image_processor _lowercase =prepare_img() _lowercase =image_processor(images=_A , return_tensors='pt' ).pixel_values.to(_A ) _lowercase =torch.tensor([[1, 2]] ) _lowercase =torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _lowercase =model( input_ids=input_ids.to(_A ) , bbox=bbox.to(_A ) , pixel_values=pixel_values.to(_A ) , ) # verify the logits _lowercase =torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , _A ) _lowercase =torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ).to(_A ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , _A , atol=1e-4 ) )
205
import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: __A : List[str] = x __A : str = y for step in range(a ): # noqa: B007 __A : Union[str, Any] = a * a - b * b + x __A : Optional[int] = 2 * a * b + y __A : List[str] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( a = 8_00 , a = 6_00 , a = -0.6 , a = 0 , a = 3.2 , a = 50 , a = True , ) -> Image.Image: __A : str = Image.new('RGB' , (image_width, image_height) ) __A : Dict = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates __A : Dict = figure_width / image_width * image_height __A : Union[str, Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width __A : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height __A : Union[str, Any] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __A : Optional[Any] = get_color_coded_rgb(a ) else: __A : Dict = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCAmelCase : str = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
280
0
'''simple docstring''' UpperCAmelCase_ = [ [0, 1_6, 1_3, 0, 0, 0], [0, 0, 1_0, 1_2, 0, 0], [0, 4, 0, 0, 1_4, 0], [0, 0, 9, 0, 0, 2_0], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ = [False] * len(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = [s] UpperCAmelCase__ = True while queue: UpperCAmelCase__ = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = True UpperCAmelCase__ = u return visited[t] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' UpperCAmelCase__ = [-1] * (len(SCREAMING_SNAKE_CASE__ )) UpperCAmelCase__ = 0 UpperCAmelCase__ = [] UpperCAmelCase__ = [i[:] for i in graph] # Record original cut, copy. while bfs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = float("""Inf""" ) UpperCAmelCase__ = sink while s != source: # Find the minimum value in select path UpperCAmelCase__ = min(SCREAMING_SNAKE_CASE__ , graph[parent[s]][s] ) UpperCAmelCase__ = parent[s] max_flow += path_flow UpperCAmelCase__ = sink while v != source: UpperCAmelCase__ = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase__ = parent[v] for i in range(len(SCREAMING_SNAKE_CASE__ ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( a , nominal_annual_percentage_rate / 3_65 , number_of_years * 3_65 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline lowerCAmelCase__ : int = logging.get_logger(__name__) @add_end_docstrings(snake_case__ ) class snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,**lowerCamelCase__ : Any ): super().__init__(**_A ) if self.framework != "pt": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) # No specific FOR_XXX available yet def __call__( self : List[str] ,lowerCamelCase__ : Any ,**lowerCamelCase__ : List[str] ): return super().__call__(_A ,**_A ) def __lowerCAmelCase ( self : Optional[int] ,**lowerCamelCase__ : List[str] ): UpperCAmelCase__ = {} if "candidate_labels" in kwargs: UpperCAmelCase__ = kwargs['candidate_labels'] if "hypothesis_template" in kwargs: UpperCAmelCase__ = kwargs['hypothesis_template'] return preprocess_params, {}, {} def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : Dict=None ,lowerCamelCase__ : str="This is a sound of {}." ): if isinstance(_A ,_A ): if audio.startswith('http://' ) or audio.startswith('https://' ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png UpperCAmelCase__ = requests.get(_A ).content else: with open(_A ,'rb' ) as f: UpperCAmelCase__ = f.read() if isinstance(_A ,_A ): UpperCAmelCase__ = ffmpeg_read(_A ,self.feature_extractor.sampling_rate ) if not isinstance(_A ,np.ndarray ): raise ValueError('We expect a numpy ndarray as input' ) if len(audio.shape ) != 1: raise ValueError('We expect a single channel audio input for ZeroShotAudioClassificationPipeline' ) UpperCAmelCase__ = self.feature_extractor( [audio] ,sampling_rate=self.feature_extractor.sampling_rate ,return_tensors='pt' ) UpperCAmelCase__ = candidate_labels UpperCAmelCase__ = [hypothesis_template.format(_A ) for x in candidate_labels] UpperCAmelCase__ = self.tokenizer(_A ,return_tensors=self.framework ,padding=_A ) UpperCAmelCase__ = [text_inputs] return inputs def __lowerCAmelCase ( self : Dict ,lowerCamelCase__ : int ): UpperCAmelCase__ = model_inputs.pop('candidate_labels' ) UpperCAmelCase__ = model_inputs.pop('text_inputs' ) if isinstance(text_inputs[0] ,_A ): UpperCAmelCase__ = text_inputs[0] else: # Batching case. UpperCAmelCase__ = text_inputs[0][0] UpperCAmelCase__ = self.model(**_A ,**_A ) UpperCAmelCase__ = { 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_audio, } return model_outputs def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : str ): UpperCAmelCase__ = model_outputs.pop('candidate_labels' ) UpperCAmelCase__ = model_outputs['logits'][0] if self.framework == "pt": UpperCAmelCase__ = logits.softmax(dim=0 ) UpperCAmelCase__ = probs.tolist() else: raise ValueError('`tf` framework not supported.' ) UpperCAmelCase__ = [ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(_A ,_A ) ,key=lambda lowerCamelCase__ : -x[0] ) ] return result
98
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_torchaudio, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class __magic_name__ ( unittest.TestCase ): """simple docstring""" __UpperCamelCase = MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING __UpperCamelCase = TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING def SCREAMING_SNAKE_CASE ( self :Tuple , snake_case :Dict , snake_case :Union[str, Any] , snake_case :str ): '''simple docstring''' A_ : Union[str, Any] = AudioClassificationPipeline(model=_A , feature_extractor=_A ) # test with a raw waveform A_ : Any = np.zeros((34_000,) ) A_ : Optional[int] = np.zeros((14_000,) ) return audio_classifier, [audioa, audio] def SCREAMING_SNAKE_CASE ( self :List[Any] , snake_case :List[Any] , snake_case :Union[str, Any] ): '''simple docstring''' A_ : int = examples A_ : int = audio_classifier(_A ) # by default a model is initialized with num_labels=2 self.assertEqual( _A , [ {"score": ANY(_A ), "label": ANY(_A )}, {"score": ANY(_A ), "label": ANY(_A )}, ] , ) A_ : Union[str, Any] = audio_classifier(_A , top_k=1 ) self.assertEqual( _A , [ {"score": ANY(_A ), "label": ANY(_A )}, ] , ) self.run_torchaudio(_A ) @require_torchaudio def SCREAMING_SNAKE_CASE ( self :List[Any] , snake_case :Any ): '''simple docstring''' import datasets # test with a local file A_ : Any = datasets.load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) A_ : Optional[Any] = dataset[0]['audio']['array'] A_ : Union[str, Any] = audio_classifier(_A ) self.assertEqual( _A , [ {"score": ANY(_A ), "label": ANY(_A )}, {"score": ANY(_A ), "label": ANY(_A )}, ] , ) @require_torch def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : Tuple = 'anton-l/wav2vec2-random-tiny-classifier' A_ : List[str] = pipeline("audio-classification" , model=_A ) A_ : Dict = np.ones((8_000,) ) A_ : int = audio_classifier(_A , top_k=4 ) A_ : List[str] = [ {'score': 0.0842, 'label': 'no'}, {'score': 0.0838, 'label': 'up'}, {'score': 0.0837, 'label': 'go'}, {'score': 0.0834, 'label': 'right'}, ] A_ : List[str] = [ {'score': 0.0845, 'label': 'stop'}, {'score': 0.0844, 'label': 'on'}, {'score': 0.0841, 'label': 'right'}, {'score': 0.0834, 'label': 'left'}, ] self.assertIn(nested_simplify(_A , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) A_ : Tuple = {'array': np.ones((8_000,) ), 'sampling_rate': audio_classifier.feature_extractor.sampling_rate} A_ : List[str] = audio_classifier(_A , top_k=4 ) self.assertIn(nested_simplify(_A , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' import datasets A_ : Union[str, Any] = 'superb/wav2vec2-base-superb-ks' A_ : Any = pipeline("audio-classification" , model=_A ) A_ : Tuple = datasets.load_dataset("anton-l/superb_dummy" , "ks" , split="test" ) A_ : Any = np.array(dataset[3]["speech"] , dtype=np.floataa ) A_ : Optional[int] = audio_classifier(_A , top_k=4 ) self.assertEqual( nested_simplify(_A , decimals=3 ) , [ {"score": 0.981, "label": "go"}, {"score": 0.007, "label": "up"}, {"score": 0.006, "label": "_unknown_"}, {"score": 0.001, "label": "down"}, ] , ) @require_tf @unittest.skip("Audio classification is not implemented for TF" ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' pass
300
def _SCREAMING_SNAKE_CASE ( a ) -> bool: return str(a ) == str(a )[::-1] def _SCREAMING_SNAKE_CASE ( a ) -> int: return int(a ) + int(str(a )[::-1] ) def _SCREAMING_SNAKE_CASE ( a = 1_00_00 ) -> int: __A : int = [] for num in range(1 , a ): __A : List[str] = 0 __A : List[Any] = num while iterations < 50: __A : str = sum_reverse(a ) iterations += 1 if is_palindrome(a ): break else: lychrel_nums.append(a ) return len(a ) if __name__ == "__main__": print(F"""{solution() = }""")
280
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { '''google/switch-base-8''': '''https://huggingface.co/google/switch-base-8/blob/main/config.json''', } class a_ ( snake_case__ ): '''simple docstring''' UpperCamelCase = '''switch_transformers''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self , A=3_2128 , A=768 , A=64 , A=2048 , A=64 , A=12 , A=3 , A=12 , A=3 , A=12 , A=8 , A=False , A=0.01 , A="float32" , A=False , A=32 , A=128 , A=0.1 , A=1e-6 , A=0.001 , A=0.001 , A=1.0 , A="relu" , A=True , A=False , A=True , A=0 , A=1 , **A , ) -> Optional[int]: _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = d_model _SCREAMING_SNAKE_CASE = d_kv _SCREAMING_SNAKE_CASE = d_ff _SCREAMING_SNAKE_CASE = num_sparse_encoder_layers _SCREAMING_SNAKE_CASE = num_layers _SCREAMING_SNAKE_CASE = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _SCREAMING_SNAKE_CASE = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: _SCREAMING_SNAKE_CASE = self.num_layers // self.num_sparse_encoder_layers else: _SCREAMING_SNAKE_CASE = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: _SCREAMING_SNAKE_CASE = self.num_decoder_layers // self.num_sparse_decoder_layers else: _SCREAMING_SNAKE_CASE = self.num_decoder_layers # HACK: this will create 0 sparse layers _SCREAMING_SNAKE_CASE = num_heads _SCREAMING_SNAKE_CASE = num_experts _SCREAMING_SNAKE_CASE = expert_capacity _SCREAMING_SNAKE_CASE = router_bias _SCREAMING_SNAKE_CASE = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f'`router_dtype` must be one of \'float32\', \'float16\' or \'bfloat16\', got {router_dtype}' ) _SCREAMING_SNAKE_CASE = router_dtype _SCREAMING_SNAKE_CASE = router_ignore_padding_tokens _SCREAMING_SNAKE_CASE = relative_attention_num_buckets _SCREAMING_SNAKE_CASE = relative_attention_max_distance _SCREAMING_SNAKE_CASE = dropout_rate _SCREAMING_SNAKE_CASE = layer_norm_epsilon _SCREAMING_SNAKE_CASE = initializer_factor _SCREAMING_SNAKE_CASE = feed_forward_proj _SCREAMING_SNAKE_CASE = use_cache _SCREAMING_SNAKE_CASE = add_router_probs _SCREAMING_SNAKE_CASE = router_z_loss_coef _SCREAMING_SNAKE_CASE = router_aux_loss_coef _SCREAMING_SNAKE_CASE = self.feed_forward_proj.split("""-""" ) _SCREAMING_SNAKE_CASE = act_info[-1] _SCREAMING_SNAKE_CASE = act_info[0] == 'gated' if len(_A ) > 1 and act_info[0] != "gated" or len(_A ) > 2: raise ValueError( f'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' """Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. """ """\'gated-gelu\' or \'relu\'""" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": _SCREAMING_SNAKE_CASE = 'gelu_new' super().__init__( pad_token_id=_A , eos_token_id=_A , is_encoder_decoder=_A , **_A , )
58
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class _A: """simple docstring""" def __init__( self , _A = None ): if components is None: __A : int = [] __A : Tuple = list(_A ) def __len__( self ): return len(self.__components ) def __str__( self ): return "(" + ",".join(map(_A , self.__components ) ) + ")" def __add__( self , _A ): __A : Optional[int] = len(self ) if size == len(_A ): __A : Any = [self.__components[i] + other.component(_A ) for i in range(_A )] return Vector(_A ) else: raise Exception('must have the same size' ) def __sub__( self , _A ): __A : Tuple = len(self ) if size == len(_A ): __A : Union[str, Any] = [self.__components[i] - other.component(_A ) for i in range(_A )] return Vector(_A ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , (float, int) ): __A : str = [c * other for c in self.__components] return Vector(_A ) elif isinstance(_A , _A ) and len(self ) == len(_A ): __A : Union[str, Any] = len(self ) __A : Dict = [self.__components[i] * other.component(_A ) for i in range(_A )] return sum(_A ) else: # error case raise Exception('invalid operand!' ) def UpperCAmelCase_ ( self ): return Vector(self.__components ) def UpperCAmelCase_ ( self , _A ): if isinstance(_A , _A ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def UpperCAmelCase_ ( self , _A , _A ): assert -len(self.__components ) <= pos < len(self.__components ) __A : Optional[int] = value def UpperCAmelCase_ ( self ): if len(self.__components ) == 0: raise Exception('Vector is empty' ) __A : Optional[Any] = [c**2 for c in self.__components] return math.sqrt(sum(_A ) ) def UpperCAmelCase_ ( self , _A , _A = False ): __A : Optional[Any] = self * other __A : Optional[Any] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def _SCREAMING_SNAKE_CASE ( a ) -> Vector: assert isinstance(a , a ) return Vector([0] * dimension ) def _SCREAMING_SNAKE_CASE ( a , a ) -> Vector: assert isinstance(a , a ) and (isinstance(a , a )) __A : Optional[Any] = [0] * dimension __A : Tuple = 1 return Vector(a ) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: assert ( isinstance(a , a ) and isinstance(a , a ) and (isinstance(a , (int, float) )) ) return x * scalar + y def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: random.seed(a ) __A : str = [random.randint(a , a ) for _ in range(a )] return Vector(a ) class _A: """simple docstring""" def __init__( self , _A , _A , _A ): __A : Optional[Any] = matrix __A : Dict = w __A : Optional[int] = h def __str__( self ): __A : Tuple = '' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Optional[Any] = [] for i in range(self.__height ): __A : Optional[Any] = [ self.__matrix[i][j] + other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Tuple = [] for i in range(self.__height ): __A : str = [ self.__matrix[i][j] - other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , _A ): # matrix-vector if len(_A ) == self.__width: __A : List[Any] = zero_vector(self.__height ) for i in range(self.__height ): __A : List[str] = [ self.__matrix[i][j] * other.component(_A ) for j in range(self.__width ) ] ans.change_component(_A , sum(_A ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(_A , (int, float) ): # matrix-scalar __A : List[str] = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(_A , self.__width , self.__height ) return None def UpperCAmelCase_ ( self ): return self.__height def UpperCAmelCase_ ( self ): return self.__width def UpperCAmelCase_ ( self , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: __A : int = value else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) __A : List[str] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(_A ) ): __A : Optional[int] = minor[i][:y] + minor[i][y + 1 :] return Matrix(_A , self.__width - 1 , self.__height - 1 ).determinant() def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(_A , _A ) else: raise Exception('Indices out of bounds' ) def UpperCAmelCase_ ( self ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __A : List[str] = [ self.__matrix[0][y] * self.cofactor(0 , _A ) for y in range(self.__width ) ] return sum(_A ) def _SCREAMING_SNAKE_CASE ( a ) -> Matrix: __A : list[list[float]] = [[0] * n for _ in range(a )] return Matrix(a , a , a ) def _SCREAMING_SNAKE_CASE ( a , a , a , a ) -> Matrix: random.seed(a ) __A : list[list[float]] = [ [random.randint(a , a ) for _ in range(a )] for _ in range(a ) ] return Matrix(a , a , a )
280
0
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @parameterized.expand([(None,), ("foo.json",)] ) def __UpperCamelCase ( self : Union[str, Any] , a : str ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = GenerationConfig( do_sample=_A , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_A , config_name=_A ) SCREAMING_SNAKE_CASE : int = GenerationConfig.from_pretrained(_A , config_name=_A ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , _A ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 50 ) self.assertEqual(loaded_config.max_length , 20 ) self.assertEqual(loaded_config.max_time , _A ) def __UpperCamelCase ( self : str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = AutoConfig.from_pretrained("gpt2" ) SCREAMING_SNAKE_CASE : str = GenerationConfig.from_model_config(_A ) SCREAMING_SNAKE_CASE : int = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(_A , _A ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def __UpperCamelCase ( self : Optional[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = GenerationConfig() SCREAMING_SNAKE_CASE : Any = { 'max_new_tokens': 1024, 'foo': 'bar', } SCREAMING_SNAKE_CASE : List[str] = copy.deepcopy(_A ) SCREAMING_SNAKE_CASE : Union[str, Any] = generation_config.update(**_A ) # update_kwargs was not modified (no side effects) self.assertEqual(_A , _A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 1024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(_A , {"foo": "bar"} ) def __UpperCamelCase ( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = GenerationConfig() SCREAMING_SNAKE_CASE : Optional[int] = 'bar' with tempfile.TemporaryDirectory("test-generation-config" ) as tmp_dir: generation_config.save_pretrained(_A ) SCREAMING_SNAKE_CASE : List[Any] = GenerationConfig.from_pretrained(_A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , "bar" ) SCREAMING_SNAKE_CASE : Any = GenerationConfig.from_model_config(_A ) assert not hasattr(_A , "foo" ) # no new kwargs should be initialized if from config def __UpperCamelCase ( self : int ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : str = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , _A ) self.assertEqual(default_config.num_beams , 1 ) SCREAMING_SNAKE_CASE : Optional[int] = GenerationConfig( do_sample=_A , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , _A ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_A ) SCREAMING_SNAKE_CASE : List[Any] = GenerationConfig.from_pretrained(_A , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , _A ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @classmethod def __UpperCamelCase ( cls : Dict ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = TOKEN HfFolder.save_token(_A ) @classmethod def __UpperCamelCase ( cls : Optional[int] ) -> str: """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-generation-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-generation-config-org" ) except HTTPError: pass def __UpperCamelCase ( self : Optional[int] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = GenerationConfig( do_sample=_A , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("test-generation-config" , use_auth_token=self._token ) SCREAMING_SNAKE_CASE : List[Any] = GenerationConfig.from_pretrained(F"{USER}/test-generation-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_A , getattr(_A , _A ) ) # Reset repo delete_repo(token=self._token , repo_id="test-generation-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( _A , repo_id="test-generation-config" , push_to_hub=_A , use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Optional[Any] = GenerationConfig.from_pretrained(F"{USER}/test-generation-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_A , getattr(_A , _A ) ) def __UpperCamelCase ( self : Dict ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = GenerationConfig( do_sample=_A , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("valid_org/test-generation-config-org" , use_auth_token=self._token ) SCREAMING_SNAKE_CASE : str = GenerationConfig.from_pretrained("valid_org/test-generation-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_A , getattr(_A , _A ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-generation-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( _A , repo_id="valid_org/test-generation-config-org" , push_to_hub=_A , use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Union[str, Any] = GenerationConfig.from_pretrained("valid_org/test-generation-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_A , getattr(_A , _A ) )
76
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase : List[str] = '''▁''' UpperCAmelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Optional[int] = BertGenerationTokenizer UpperCamelCase : str = False UpperCamelCase : Tuple = True def UpperCAmelCase_ ( self ): super().setUp() __A : Tuple = BertGenerationTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : str = '<s>' __A : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def UpperCAmelCase_ ( self ): __A : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(_A ) , 1002 ) def UpperCAmelCase_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCAmelCase_ ( self ): __A : str = BertGenerationTokenizer(_A , keep_accents=_A ) __A : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_A , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [285, 46, 10, 170, 382] , ) __A : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __A : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __A : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase_ ( self ): return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def UpperCAmelCase_ ( self ): __A : List[Any] = 'Hello World!' __A : Optional[Any] = [18536, 2260, 101] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def UpperCAmelCase_ ( self ): __A : Dict = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) __A : int = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @require_torch @slow def UpperCAmelCase_ ( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence __A : Tuple = list(self.big_tokenizer.get_vocab().keys() )[:10] __A : List[Any] = ' '.join(_A ) __A : Union[str, Any] = self.big_tokenizer.encode_plus(_A , return_tensors='pt' , return_token_type_ids=_A ) __A : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=_A ) __A : int = BertGenerationConfig() __A : List[str] = BertGenerationEncoder(_A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def UpperCAmelCase_ ( self ): # fmt: off __A : str = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
280
0
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict UpperCamelCase__ = namedtuple( '_TestCommandArgs', [ 'dataset', 'name', 'cache_dir', 'data_dir', 'all_configs', 'save_infos', 'ignore_verifications', 'force_redownload', 'clear_cache', ], defaults=[None, None, None, False, False, False, False, False], ) def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = _TestCommandArgs(dataset=__A, all_configs=__A, save_infos=__A ) UpperCAmelCase__ = TestCommand(*__A ) test_command.run() UpperCAmelCase__ = os.path.join(__A, "README.md" ) assert os.path.exists(__A ) UpperCAmelCase__ = DatasetInfosDict.from_directory(__A ) UpperCAmelCase__ = DatasetInfosDict( { "default": DatasetInfo( features=Features( { "tokens": Sequence(Value("string" ) ), "ner_tags": Sequence( ClassLabel(names=["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] ) ), "langs": Sequence(Value("string" ) ), "spans": Sequence(Value("string" ) ), } ), splits=[ { "name": "train", "num_bytes": 2_351_563, "num_examples": 10_000, }, { "name": "validation", "num_bytes": 238_418, "num_examples": 1_000, }, ], download_size=3_940_680, dataset_size=2_589_981, ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: UpperCAmelCase__ = getattr(dataset_infos["default"], __A ), getattr(expected_dataset_infos["default"], __A ) if key == "num_bytes": assert is_apercent_close(__A, __A ) elif key == "splits": assert list(__A ) == list(__A ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes, expected[split].num_bytes ) else: result == expected
65
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _A: """simple docstring""" @staticmethod def UpperCAmelCase_ ( *_A , **_A ): pass def _SCREAMING_SNAKE_CASE ( a ) -> str: __A : str = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def _SCREAMING_SNAKE_CASE ( a ) -> Dict: __A : Dict = np.array(a ) __A : List[Any] = npimg.shape return {"hash": hashimage(a ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _A( unittest.TestCase ): """simple docstring""" UpperCamelCase : str = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase : int = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def UpperCAmelCase_ ( self , _A , _A , _A ): __A : Dict = MaskGenerationPipeline(model=_A , image_processor=_A ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase_ ( self , _A , _A ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def UpperCAmelCase_ ( self ): pass @slow @require_torch def UpperCAmelCase_ ( self ): __A : Union[str, Any] = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) __A : List[str] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=256 ) # Shortening by hashing __A : List[Any] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] , ) # fmt: on @require_torch @slow def UpperCAmelCase_ ( self ): __A : Optional[Any] = 'facebook/sam-vit-huge' __A : List[str] = pipeline('mask-generation' , model=_A ) __A : Tuple = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __A : List[str] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] , )
280
0
"""simple docstring""" import pandas as pd from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split # Fitting Polynomial Regression to the dataset from sklearn.preprocessing import PolynomialFeatures # Importing the dataset lowercase__ = pd.read_csv( 'https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/' 'position_salaries.csv' ) lowercase__ = dataset.iloc[:, 1:2].values lowercase__ = dataset.iloc[:, 2].values lowercase__ = train_test_split(X, y, test_size=0.2, random_state=0) lowercase__ = PolynomialFeatures(degree=4) lowercase__ = poly_reg.fit_transform(X) lowercase__ = LinearRegression() pol_reg.fit(X_poly, y) def __a ( ) ->Tuple: plt.scatter(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , color='red' ) plt.plot(_SCREAMING_SNAKE_CASE , pol_reg.predict(poly_reg.fit_transform(_SCREAMING_SNAKE_CASE ) ) , color='blue' ) plt.title('Truth or Bluff (Linear Regression)' ) plt.xlabel('Position level' ) plt.ylabel('Salary' ) plt.show() if __name__ == "__main__": viz_polymonial() # Predicting a new result with Polymonial Regression pol_reg.predict(poly_reg.fit_transform([[5.5]])) # output should be 132148.43750003
290
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): __A : List[Any] = tempfile.mkdtemp() # fmt: off __A : List[str] = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on __A : Union[str, Any] = dict(zip(_A , range(len(_A ) ) ) ) __A : Optional[int] = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] __A : int = {'unk_token': '<unk>'} __A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_A ) ) __A : List[Any] = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], 'image_std': [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } __A : Optional[int] = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def UpperCAmelCase_ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __A : Optional[int] = [Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase_ ( self ): __A : List[Any] = self.get_tokenizer() __A : str = self.get_rust_tokenizer() __A : List[str] = self.get_image_processor() __A : Optional[int] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_slow.save_pretrained(self.tmpdirname ) __A : int = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_fast.save_pretrained(self.tmpdirname ) __A : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _A ) self.assertIsInstance(processor_fast.tokenizer , _A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _A ) self.assertIsInstance(processor_fast.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : List[str] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) __A : Optional[int] = self.get_image_processor(do_normalize=_A ) __A : Any = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Optional[Any] = self.get_tokenizer() __A : Union[str, Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Union[str, Any] = self.prepare_image_inputs() __A : int = image_processor(_A , return_tensors='np' ) __A : str = processor(images=_A , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCAmelCase_ ( self ): __A : str = self.get_image_processor() __A : str = self.get_tokenizer() __A : Tuple = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : str = 'lower newer' __A : str = processor(text=_A , return_tensors='np' ) __A : List[str] = tokenizer(_A , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCAmelCase_ ( self ): __A : int = self.get_image_processor() __A : Optional[int] = self.get_tokenizer() __A : List[str] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Any = 'lower newer' __A : Optional[Any] = self.prepare_image_inputs() __A : List[Any] = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Any = 'google/owlvit-base-patch32' __A : int = OwlViTProcessor.from_pretrained(_A ) __A : Dict = ['cat', 'nasa badge'] __A : Optional[Any] = processor(text=_A ) __A : Optional[int] = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Tuple = 'google/owlvit-base-patch32' __A : Any = OwlViTProcessor.from_pretrained(_A ) __A : Dict = [['cat', 'nasa badge'], ['person']] __A : Dict = processor(text=_A ) __A : Optional[int] = 16 __A : Any = len(_A ) __A : Union[str, Any] = max([len(_A ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : List[Any] = 'google/owlvit-base-patch32' __A : str = OwlViTProcessor.from_pretrained(_A ) __A : Union[str, Any] = ['cat', 'nasa badge'] __A : Tuple = processor(text=_A ) __A : str = 16 __A : int = inputs['input_ids'] __A : List[Any] = [ [49406, 2368, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49406, 6841, 11301, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : List[str] = self.get_tokenizer() __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = processor(images=_A , query_images=_A ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Union[str, Any] = self.get_tokenizer() __A : str = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A : Any = processor.batch_decode(_A ) __A : Tuple = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A )
280
0
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE__ : List[str] = '''▁''' SCREAMING_SNAKE_CASE__ : Optional[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class UpperCamelCase__ (snake_case__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Optional[int] = BertGenerationTokenizer lowerCamelCase_ : str = False lowerCamelCase_ : Tuple = True def _lowercase ( self ) -> Optional[Any]: super().setUp() lowerCamelCase : Tuple = BertGenerationTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self ) -> List[str]: lowerCamelCase : str = '<s>' lowerCamelCase : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def _lowercase ( self ) -> Optional[int]: lowerCamelCase : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "<pad>" ) self.assertEqual(len(_A ) , 1002 ) def _lowercase ( self ) -> List[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def _lowercase ( self ) -> Tuple: lowerCamelCase : str = BertGenerationTokenizer(_A , keep_accents=_A ) lowerCamelCase : Dict = tokenizer.tokenize("This is a test" ) self.assertListEqual(_A , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [285, 46, 10, 170, 382] , ) lowerCamelCase : int = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) lowerCamelCase : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowerCamelCase : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def _lowercase ( self ) -> List[str]: return BertGenerationTokenizer.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) @slow def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : List[Any] = 'Hello World!' lowerCamelCase : Optional[Any] = [1_8536, 2260, 101] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def _lowercase ( self ) -> Optional[int]: lowerCamelCase : Dict = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCamelCase : int = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 3_4324, 497, 391, 408, 1_1342, 1244, 385, 100, 938, 985, 456, 574, 362, 1_2597, 3200, 3129, 1172, ] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @require_torch @slow def _lowercase ( self ) -> str: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowerCamelCase : Tuple = list(self.big_tokenizer.get_vocab().keys() )[:10] lowerCamelCase : List[Any] = ' '.join(_A ) lowerCamelCase : Union[str, Any] = self.big_tokenizer.encode_plus(_A , return_tensors="pt" , return_token_type_ids=_A ) lowerCamelCase : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=_A ) lowerCamelCase : int = BertGenerationConfig() lowerCamelCase : List[str] = BertGenerationEncoder(_A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def _lowercase ( self ) -> Union[str, Any]: # fmt: off lowerCamelCase : str = {'input_ids': [[3_9286, 458, 3_6335, 2001, 456, 1_3073, 1_3266, 455, 113, 7746, 1741, 1_1157, 391, 1_3073, 1_3266, 455, 113, 3967, 3_5412, 113, 4936, 109, 3870, 2377, 113, 3_0084, 4_5720, 458, 134, 1_7496, 112, 503, 1_1672, 113, 118, 112, 5665, 1_3347, 3_8687, 112, 1496, 3_1389, 112, 3268, 4_7264, 134, 962, 112, 1_6377, 8035, 2_3130, 430, 1_2169, 1_5518, 2_8592, 458, 146, 4_1697, 109, 391, 1_2169, 1_5518, 1_6689, 458, 146, 4_1358, 109, 452, 726, 4034, 111, 763, 3_5412, 5082, 388, 1903, 111, 9051, 391, 2870, 4_8918, 1900, 1123, 550, 998, 112, 9586, 1_5985, 455, 391, 410, 2_2955, 3_7636, 114], [448, 1_7496, 419, 3663, 385, 763, 113, 2_7533, 2870, 3283, 1_3043, 1639, 2_4713, 523, 656, 2_4013, 1_8550, 2521, 517, 2_7014, 2_1244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 1_1786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 2_1932, 1_8146, 726, 363, 1_7032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name="google/bert_for_seq_generation_L-24_bbc_encoder" , revision="c817d1fd1be2ffa69431227a1fe320544943d4db" , )
48
import math def _SCREAMING_SNAKE_CASE ( a ) -> list[int]: __A : List[str] = [] __A : Any = 2 __A : Union[str, Any] = int(math.sqrt(a ) ) # Size of every segment __A : Any = [True] * (end + 1) __A : List[Any] = [] while start <= end: if temp[start] is True: in_prime.append(a ) for i in range(start * start , end + 1 , a ): __A : Optional[int] = False start += 1 prime += in_prime __A : Any = end + 1 __A : Any = min(2 * end , a ) while low <= n: __A : List[Any] = [True] * (high - low + 1) for each in in_prime: __A : List[str] = math.floor(low / each ) * each if t < low: t += each for j in range(a , high + 1 , a ): __A : Optional[int] = False for j in range(len(a ) ): if temp[j] is True: prime.append(j + low ) __A : Optional[int] = high + 1 __A : Tuple = min(high + end , a ) return prime print(sieve(10**6))
280
0
from __future__ import annotations def __UpperCAmelCase ( a_ , a_ , a_ , ): if (stress, tangential_force, area).count(0) != 1: raise ValueError('You cannot supply more or less than 2 values') elif stress < 0: raise ValueError('Stress cannot be negative') elif tangential_force < 0: raise ValueError('Tangential Force cannot be negative') elif area < 0: raise ValueError('Area cannot be negative') elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
178
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : str = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase_ = {'''configuration_glpn''': ['''GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GLPNConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ['''GLPNFeatureExtractor'''] UpperCamelCase_ = ['''GLPNImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ '''GLPN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GLPNForDepthEstimation''', '''GLPNLayer''', '''GLPNModel''', '''GLPNPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
309
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract lowercase_ = logging.get_logger(__name__) def a ( A__ : Tuple , A__ : Union[str, Any] , A__ : List[str] ) -> Any: """simple docstring""" return [ int(1000 * (box[0] / width) ), int(1000 * (box[1] / height) ), int(1000 * (box[2] / width) ), int(1000 * (box[3] / height) ), ] def a ( A__ : Optional[int] , A__ : List[Any] , A__ : int = None ) -> Optional[int]: """simple docstring""" _lowercase =tesseract_config if tesseract_config is not None else '' # apply OCR _lowercase =to_pil_image(A__ ) _lowercase =pil_image.size _lowercase =pytesseract.image_to_data(A__ , lang=A__ , output_type='dict' , config=A__ ) _lowercase =data['text'], data['left'], data['top'], data['width'], data['height'] # filter empty words and corresponding coordinates _lowercase =[idx for idx, word in enumerate(A__ ) if not word.strip()] _lowercase =[word for idx, word in enumerate(A__ ) if idx not in irrelevant_indices] _lowercase =[coord for idx, coord in enumerate(A__ ) if idx not in irrelevant_indices] _lowercase =[coord for idx, coord in enumerate(A__ ) if idx not in irrelevant_indices] _lowercase =[coord for idx, coord in enumerate(A__ ) if idx not in irrelevant_indices] _lowercase =[coord for idx, coord in enumerate(A__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format _lowercase =[] for x, y, w, h in zip(A__ , A__ , A__ , A__ ): _lowercase =[x, y, x + w, y + h] actual_boxes.append(A__ ) # finally, normalize the bounding boxes _lowercase =[] for box in actual_boxes: normalized_boxes.append(normalize_box(A__ , A__ , A__ ) ) assert len(A__ ) == len(A__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class __lowerCAmelCase ( snake_case__ ): _a = ['''pixel_values'''] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = "" , **lowerCAmelCase , ) -> List[str]: '''simple docstring''' super().__init__(**_A ) _lowercase =size if size is not None else {'height': 224, 'width': 224} _lowercase =get_size_dict(_A ) _lowercase =do_resize _lowercase =size _lowercase =resample _lowercase =apply_ocr _lowercase =ocr_lang _lowercase =tesseract_config def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ) -> Dict: '''simple docstring''' _lowercase =get_size_dict(_A ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) _lowercase =(size['height'], size['width']) return resize(_A , size=_A , resample=_A , data_format=_A , **_A ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ) -> Any: '''simple docstring''' _lowercase =do_resize if do_resize is not None else self.do_resize _lowercase =size if size is not None else self.size _lowercase =get_size_dict(_A ) _lowercase =resample if resample is not None else self.resample _lowercase =apply_ocr if apply_ocr is not None else self.apply_ocr _lowercase =ocr_lang if ocr_lang is not None else self.ocr_lang _lowercase =tesseract_config if tesseract_config is not None else self.tesseract_config _lowercase =make_list_of_images(_A ) if not valid_images(_A ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) # All transformations expect numpy arrays. _lowercase =[to_numpy_array(_A ) for image in images] if apply_ocr: requires_backends(self , 'pytesseract' ) _lowercase =[] _lowercase =[] for image in images: _lowercase =apply_tesseract(_A , _A , _A ) words_batch.append(_A ) boxes_batch.append(_A ) if do_resize: _lowercase =[self.resize(image=_A , size=_A , resample=_A ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) _lowercase =[flip_channel_order(_A ) for image in images] _lowercase =[to_channel_dimension_format(_A , _A ) for image in images] _lowercase =BatchFeature(data={'pixel_values': images} , tensor_type=_A ) if apply_ocr: _lowercase =words_batch _lowercase =boxes_batch return data
205
def _SCREAMING_SNAKE_CASE ( a , a = 0 ) -> list: __A : int = length or len(a ) __A : str = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __A , __A : Optional[int] = list_data[i + 1], list_data[i] __A : Union[str, Any] = True return list_data if not swapped else bubble_sort(a , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' from __future__ import annotations import math def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Dict ): '''simple docstring''' UpperCAmelCase__ = u for i in range(1 , SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = temp * (u - i) return temp def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = int(input("""enter the numbers of values: """ ) ) UpperCAmelCase__ = [] for _ in range(SCREAMING_SNAKE_CASE__ ): y.append([] ) for i in range(SCREAMING_SNAKE_CASE__ ): for j in range(SCREAMING_SNAKE_CASE__ ): y[i].append(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = 0 print("""enter the values of parameters in a list: """ ) UpperCAmelCase__ = list(map(SCREAMING_SNAKE_CASE__ , input().split() ) ) print("""enter the values of corresponding parameters: """ ) for i in range(SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = float(input() ) UpperCAmelCase__ = int(input("""enter the value to interpolate: """ ) ) UpperCAmelCase__ = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , SCREAMING_SNAKE_CASE__ ): for j in range(n - i ): UpperCAmelCase__ = y[j + 1][i - 1] - y[j][i - 1] UpperCAmelCase__ = y[0][0] for i in range(1 , SCREAMING_SNAKE_CASE__ ): summ += (ucal(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) * y[0][i]) / math.factorial(SCREAMING_SNAKE_CASE__ ) print(F'''the value at {value} is {summ}''' ) if __name__ == "__main__": main()
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a ) -> int: if not nums: return 0 __A : Optional[int] = nums[0] __A : str = 0 for num in nums[1:]: __A , __A : Tuple = ( max_excluding + num, max(a , a ), ) return max(a , a ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = int(lowerCamelCase ) # Initialize Result UpperCAmelCase__ = [] # Traverse through all denomination for denomination in reversed(lowerCamelCase ): # Find denominations while int(lowerCamelCase ) >= int(lowerCamelCase ): total_value -= int(lowerCamelCase ) answer.append(lowerCamelCase ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : Optional[int] = '''0''' if ( input('Do you want to enter your denominations ? (yY/n): ').strip().lower() == "y" ): lowerCAmelCase__ : List[Any] = int(input('Enter the number of denominations you want to add: ').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) lowerCAmelCase__ : int = input('Enter the change you want to make in Indian Currency: ').strip() else: # All denominations of Indian Currency if user does not enter lowerCAmelCase__ : Optional[int] = [1, 2, 5, 10, 20, 50, 100, 500, 2_000] lowerCAmelCase__ : Tuple = input('Enter the change you want to make: ').strip() if int(value) == 0 or int(value) < 0: print('The total value cannot be zero or negative.') else: print(F"""Following is minimal change for {value}: """) lowerCAmelCase__ : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=' ')
98
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class __magic_name__ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : str = 'laion/clap-htsat-unfused' A_ : Dict = tempfile.mkdtemp() def SCREAMING_SNAKE_CASE ( self :str , **snake_case :str ): '''simple docstring''' return RobertaTokenizer.from_pretrained(self.checkpoint , **_A ) def SCREAMING_SNAKE_CASE ( self :Any , **snake_case :List[Any] ): '''simple docstring''' return ClapFeatureExtractor.from_pretrained(self.checkpoint , **_A ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : Tuple = self.get_tokenizer() A_ : Any = self.get_feature_extractor() A_ : Union[str, Any] = ClapProcessor(tokenizer=_A , feature_extractor=_A ) processor.save_pretrained(self.tmpdirname ) A_ : Any = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : str = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) A_ : Optional[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) A_ : List[Any] = self.get_feature_extractor(do_normalize=_A , padding_value=1.0 ) A_ : Any = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=_A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : Tuple = self.get_feature_extractor() A_ : Optional[Any] = self.get_tokenizer() A_ : Optional[int] = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ : int = floats_list((3, 1_000) ) A_ : List[Any] = feature_extractor(_A , return_tensors="np" ) A_ : List[str] = processor(audios=_A , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Any = self.get_feature_extractor() A_ : str = self.get_tokenizer() A_ : Dict = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ : Optional[Any] = 'This is a test string' A_ : Any = processor(text=_A ) A_ : Optional[int] = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : List[Any] = self.get_feature_extractor() A_ : Union[str, Any] = self.get_tokenizer() A_ : str = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A_ : Dict = processor.batch_decode(_A ) A_ : List[Any] = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : List[str] = self.get_feature_extractor() A_ : Any = self.get_tokenizer() A_ : Any = ClapProcessor(tokenizer=_A , feature_extractor=_A ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , )
300
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def lowerCamelCase ( __lowerCamelCase : Dict ) ->Tuple: print("""Loading config file...""" ) def flatten_yaml_as_dict(__lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any]="" , __lowerCamelCase : List[Any]="." ): _SCREAMING_SNAKE_CASE = [] for k, v in d.items(): _SCREAMING_SNAKE_CASE = parent_key + sep + k if parent_key else k if isinstance(__lowerCamelCase , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(__lowerCamelCase , __lowerCamelCase , sep=__lowerCamelCase ).items() ) else: items.append((new_key, v) ) return dict(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = argparse.Namespace() with open(__lowerCamelCase , """r""" ) as yaml_file: try: _SCREAMING_SNAKE_CASE = yaml.load(__lowerCamelCase , Loader=yaml.FullLoader ) _SCREAMING_SNAKE_CASE = flatten_yaml_as_dict(__lowerCamelCase ) for k, v in flat_cfg.items(): setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) except yaml.YAMLError as exc: logger.error("""Error while loading config file: {}. Error message: {}""".format(__lowerCamelCase , str(__lowerCamelCase ) ) ) return config def lowerCamelCase ( __lowerCamelCase : Tuple , __lowerCamelCase : Union[str, Any] ) ->List[str]: _SCREAMING_SNAKE_CASE = MobileViTVaConfig() _SCREAMING_SNAKE_CASE = False # dataset if task_name.startswith("""imagenet1k_""" ): _SCREAMING_SNAKE_CASE = 1000 if int(task_name.strip().split("""_""" )[-1] ) == 384: _SCREAMING_SNAKE_CASE = 384 else: _SCREAMING_SNAKE_CASE = 256 _SCREAMING_SNAKE_CASE = 'imagenet-1k-id2label.json' elif task_name.startswith("""imagenet21k_to_1k_""" ): _SCREAMING_SNAKE_CASE = 2_1000 if int(task_name.strip().split("""_""" )[-1] ) == 384: _SCREAMING_SNAKE_CASE = 384 else: _SCREAMING_SNAKE_CASE = 256 _SCREAMING_SNAKE_CASE = 'imagenet-22k-id2label.json' elif task_name.startswith("""ade20k_""" ): _SCREAMING_SNAKE_CASE = 151 _SCREAMING_SNAKE_CASE = 512 _SCREAMING_SNAKE_CASE = 'ade20k-id2label.json' _SCREAMING_SNAKE_CASE = True elif task_name.startswith("""voc_""" ): _SCREAMING_SNAKE_CASE = 21 _SCREAMING_SNAKE_CASE = 512 _SCREAMING_SNAKE_CASE = 'pascal-voc-id2label.json' _SCREAMING_SNAKE_CASE = True # orig_config _SCREAMING_SNAKE_CASE = load_orig_config_file(__lowerCamelCase ) assert getattr(__lowerCamelCase , """model.classification.name""" , -1 ) == "mobilevit_v2", "Invalid model" _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.classification.mitv2.width_multiplier""" , 1.0 ) assert ( getattr(__lowerCamelCase , """model.classification.mitv2.attn_norm_layer""" , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.classification.activation.name""" , """swish""" ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.segmentation.output_stride""" , 16 ) if "_deeplabv3" in task_name: _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.segmentation.deeplabv3.aspp_rates""" , [12, 24, 36] ) _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.segmentation.deeplabv3.aspp_out_channels""" , 512 ) _SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , """model.segmentation.deeplabv3.aspp_dropout""" , 0.1 ) # id2label _SCREAMING_SNAKE_CASE = 'huggingface/label-files' _SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="""dataset""" ) , """r""" ) ) _SCREAMING_SNAKE_CASE = {int(__lowerCamelCase ): v for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE = idalabel _SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config def lowerCamelCase ( __lowerCamelCase : List[str] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ) ->Any: _SCREAMING_SNAKE_CASE = dct.pop(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = val def lowerCamelCase ( __lowerCamelCase : int , __lowerCamelCase : Dict=False ) ->Optional[int]: if base_model: _SCREAMING_SNAKE_CASE = '' else: _SCREAMING_SNAKE_CASE = 'mobilevitv2.' _SCREAMING_SNAKE_CASE = [] for k in state_dict.keys(): if k[:8] == "encoder.": _SCREAMING_SNAKE_CASE = k[8:] else: _SCREAMING_SNAKE_CASE = k if ".block." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".block.""" , """.""" ) if ".conv." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".conv.""" , """.convolution.""" ) if ".norm." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".norm.""" , """.normalization.""" ) if "conv_1." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""conv_1.""" , F'{model_prefix}conv_stem.' ) for i in [1, 2]: if F'layer_{i}.' in k: _SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.' , F'{model_prefix}encoder.layer.{i-1}.layer.' ) if ".exp_1x1." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".exp_1x1.""" , """.expand_1x1.""" ) if ".red_1x1." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".red_1x1.""" , """.reduce_1x1.""" ) for i in [3, 4, 5]: if F'layer_{i}.0.' in k: _SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.0.' , F'{model_prefix}encoder.layer.{i-1}.downsampling_layer.' ) if F'layer_{i}.1.local_rep.0.' in k: _SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.local_rep.0.' , F'{model_prefix}encoder.layer.{i-1}.conv_kxk.' ) if F'layer_{i}.1.local_rep.1.' in k: _SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.local_rep.1.' , F'{model_prefix}encoder.layer.{i-1}.conv_1x1.' ) for i in [3, 4, 5]: if i == 3: _SCREAMING_SNAKE_CASE = [0, 1] elif i == 4: _SCREAMING_SNAKE_CASE = [0, 1, 2, 3] elif i == 5: _SCREAMING_SNAKE_CASE = [0, 1, 2] for j in j_in: if F'layer_{i}.1.global_rep.{j}.' in k: _SCREAMING_SNAKE_CASE = k_new.replace( F'layer_{i}.1.global_rep.{j}.' , F'{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.' ) if F'layer_{i}.1.global_rep.{j+1}.' in k: _SCREAMING_SNAKE_CASE = k_new.replace( F'layer_{i}.1.global_rep.{j+1}.' , F'{model_prefix}encoder.layer.{i-1}.layernorm.' ) if F'layer_{i}.1.conv_proj.' in k: _SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.conv_proj.' , F'{model_prefix}encoder.layer.{i-1}.conv_projection.' ) if "pre_norm_attn.0." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""pre_norm_attn.0.""" , """layernorm_before.""" ) if "pre_norm_attn.1." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""pre_norm_attn.1.""" , """attention.""" ) if "pre_norm_ffn.0." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""pre_norm_ffn.0.""" , """layernorm_after.""" ) if "pre_norm_ffn.1." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""pre_norm_ffn.1.""" , """ffn.conv1.""" ) if "pre_norm_ffn.3." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""pre_norm_ffn.3.""" , """ffn.conv2.""" ) if "classifier.1." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""classifier.1.""" , """classifier.""" ) if "seg_head." in k: _SCREAMING_SNAKE_CASE = k_new.replace("""seg_head.""" , """segmentation_head.""" ) if ".aspp_layer." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".aspp_layer.""" , """.""" ) if ".aspp_pool." in k: _SCREAMING_SNAKE_CASE = k_new.replace(""".aspp_pool.""" , """.""" ) rename_keys.append((k, k_new) ) return rename_keys def lowerCamelCase ( __lowerCamelCase : Any ) ->Optional[Any]: _SCREAMING_SNAKE_CASE = [] for k in state_dict.keys(): if k.startswith("""seg_head.aux_head.""" ): keys_to_ignore.append(__lowerCamelCase ) for k in keys_to_ignore: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase ( ) ->List[str]: _SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" _SCREAMING_SNAKE_CASE = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def lowerCamelCase ( __lowerCamelCase : int , __lowerCamelCase : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any] ) ->Union[str, Any]: _SCREAMING_SNAKE_CASE = get_mobilevitva_config(__lowerCamelCase , __lowerCamelCase ) # load original state_dict _SCREAMING_SNAKE_CASE = torch.load(__lowerCamelCase , map_location="""cpu""" ) # load huggingface model if task_name.startswith("""ade20k_""" ) or task_name.startswith("""voc_""" ): _SCREAMING_SNAKE_CASE = MobileViTVaForSemanticSegmentation(__lowerCamelCase ).eval() _SCREAMING_SNAKE_CASE = False else: _SCREAMING_SNAKE_CASE = MobileViTVaForImageClassification(__lowerCamelCase ).eval() _SCREAMING_SNAKE_CASE = False # remove and rename some keys of load the original model _SCREAMING_SNAKE_CASE = checkpoint remove_unused_keys(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = create_rename_keys(__lowerCamelCase , base_model=__lowerCamelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # load modified state_dict model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by MobileViTImageProcessor _SCREAMING_SNAKE_CASE = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) _SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = model(**__lowerCamelCase ) # verify classification model if task_name.startswith("""imagenet""" ): _SCREAMING_SNAKE_CASE = outputs.logits _SCREAMING_SNAKE_CASE = logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) if task_name.startswith("""imagenet1k_256""" ) and config.width_multiplier == 1.0: # expected_logits for base variant _SCREAMING_SNAKE_CASE = torch.tensor([-1.6_3_3_6e0_0, -7.3_2_0_4e-0_2, -5.1_8_8_3e-0_1] ) assert torch.allclose(logits[0, :3] , __lowerCamelCase , atol=1e-4 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'Saving model {task_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) lowercase_ = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
58
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> None: __A : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a ) == len(a ), F"""{len(a )} != {len(a )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) UpperCAmelCase : List[Any] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } UpperCAmelCase : Optional[int] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _SCREAMING_SNAKE_CASE ( a , a ) -> Dict: try: __A : int = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a ) ) def _SCREAMING_SNAKE_CASE ( a , a ) -> List[int]: if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _SCREAMING_SNAKE_CASE ( a , a = "student" , a = None , a = None , a=False , a=None , a=None , **a , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __A : List[str] = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(a , a ): AutoTokenizer.from_pretrained(a ).save_pretrained(a ) # purely for convenience __A : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained(a ).eval() else: assert isinstance(a , a ), F"""teacher must be a model or string got type {type(a )}""" __A : int = teacher.config.to_diff_dict() try: __A , __A : List[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A : str = teacher_e if d is None: __A : List[Any] = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): __A , __A : List[Any] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A : Optional[int] = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A : int = teacher_e if d is None: __A : Optional[Any] = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a ) # Copy weights __A : Dict = teacher.config_class(**a ) __A : int = AutoModelForSeqaSeqLM.from_config(a ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A : Any = student.load_state_dict(teacher.state_dict() , strict=a ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A : Optional[int] = list(range(a ) ), list(range(a ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) if d_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) try: if hasattr( a , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a ) copy_layers(teacher.decoder.block , student.decoder.block , a ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __A : Optional[int] = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(a ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
280
0
def lowerCamelCase__ ( _a): return " ".join( "".join(word[::-1]) if len(_a) > 4 else word for word in sentence.split()) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('Hey wollef sroirraw'))
76
def _SCREAMING_SNAKE_CASE ( a , a ) -> list[int]: __A : Optional[int] = int(a ) # Initialize Result __A : Optional[int] = [] # Traverse through all denomination for denomination in reversed(a ): # Find denominations while int(a ) >= int(a ): total_value -= int(a ) answer.append(a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase : List[str] = [] UpperCAmelCase : Optional[int] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): UpperCAmelCase : List[Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) UpperCAmelCase : int = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCAmelCase : Tuple = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(F"""Following is minimal change for {value}: """) UpperCAmelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
280
0
def lowerCAmelCase_ ( __A, __A ) -> Optional[Any]: '''simple docstring''' if b == 0: return 1 if (b % 2) == 0: return actual_power(__A, int(b / 2 ) ) * actual_power(__A, int(b / 2 ) ) else: return a * actual_power(__A, int(b / 2 ) ) * actual_power(__A, int(b / 2 ) ) def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' if b < 0: return 1 / actual_power(__A, __A ) return actual_power(__A, __A ) if __name__ == "__main__": print(power(-2, -3))
65
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _A( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __A : List[Any] = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __A : Union[str, Any] = parent __A : Optional[int] = batch_size __A : int = num_channels __A : int = min_resolution __A : Any = max_resolution __A : List[Any] = do_resize __A : List[Any] = size __A : Union[str, Any] = do_normalize __A : Optional[int] = image_mean __A : Optional[int] = image_std __A : int = do_rescale __A : str = rescale_factor __A : Tuple = do_pad def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase_ ( self , _A , _A=False ): if not batched: __A : List[str] = image_inputs[0] if isinstance(_A , Image.Image ): __A , __A : int = image.size else: __A , __A : Any = image.shape[1], image.shape[2] if w < h: __A : List[Any] = int(self.size['shortest_edge'] * h / w ) __A : List[Any] = self.size['shortest_edge'] elif w > h: __A : Union[str, Any] = self.size['shortest_edge'] __A : str = int(self.size['shortest_edge'] * w / h ) else: __A : Dict = self.size['shortest_edge'] __A : str = self.size['shortest_edge'] else: __A : int = [] for image in image_inputs: __A , __A : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __A : List[str] = max(_A , key=lambda _A : item[0] )[0] __A : str = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : List[str] = YolosImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): __A : Dict = YolosImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): __A : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) def UpperCAmelCase_ ( self ): __A : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , _A ) __A : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _A ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing __A : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input __A : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A , __A : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) __A : str = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input __A : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : List[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Tuple = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input __A : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Union[str, Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Optional[int] = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processings __A : Tuple = self.image_processing_class(**self.image_processor_dict ) __A : Any = self.image_processing_class(do_resize=_A , do_normalize=_A , do_rescale=_A ) # create random PyTorch tensors __A : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __A : Optional[int] = image_processing_a.pad(_A , return_tensors='pt' ) __A : Optional[int] = image_processing_a(_A , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1e-4 ) ) @slow def UpperCAmelCase_ ( self ): # prepare image and target __A : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __A : Optional[Any] = json.loads(f.read() ) __A : Optional[Any] = {'image_id': 39769, 'annotations': target} # encode them __A : str = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) __A : List[Any] = image_processing(images=_A , annotations=_A , return_tensors='pt' ) # verify pixel values __A : List[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : List[Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Any = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify orig_size __A : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : str = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) ) @slow def UpperCAmelCase_ ( self ): # prepare image, target and masks_path __A : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __A : Tuple = json.loads(f.read() ) __A : Any = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __A : List[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __A : Any = YolosImageProcessor(format='coco_panoptic' ) __A : List[Any] = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='pt' ) # verify pixel values __A : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Union[str, Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : List[str] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify masks __A : Tuple = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _A ) # verify orig_size __A : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : int = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) )
280
0
"""simple docstring""" from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowercase__ = logging.get_logger(__name__) class __snake_case ( snake_case__ ): def lowerCamelCase_ ( self , lowercase) -> Union[str, Any]: '''simple docstring''' if isinstance(_A , _A): a__: Optional[Any] = [label.strip() for label in labels.split(',') if label.strip()] return labels def __call__( self , lowercase , lowercase , lowercase) -> Optional[int]: '''simple docstring''' if len(_A) == 0 or len(_A) == 0: raise ValueError('You must include at least one label and at least one sequence.') if hypothesis_template.format(labels[0]) == hypothesis_template: raise ValueError( ( 'The provided hypothesis_template "{}" was not able to be formatted with the target labels. ' 'Make sure the passed template includes formatting syntax such as {{}} where the label should go.' ).format(_A)) if isinstance(_A , _A): a__: Optional[int] = [sequences] a__: Any = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(_A)] for label in labels]) return sequence_pairs, sequences @add_end_docstrings(snake_case__ ) class __snake_case ( snake_case__ ): def __init__( self , lowercase=ZeroShotClassificationArgumentHandler() , *lowercase , **lowercase) -> List[Any]: '''simple docstring''' a__: Optional[int] = args_parser super().__init__(*_A , **_A) if self.entailment_id == -1: logger.warning( 'Failed to determine \'entailment\' label id from the label2id mapping in the model config. Setting to ' '-1. Define a descriptive label2id mapping in the model config to ensure correct outputs.') @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith('entail'): return ind return -1 def lowerCamelCase_ ( self , lowercase , lowercase=True , lowercase=True , lowercase=TruncationStrategy.ONLY_FIRST , **lowercase) -> Any: '''simple docstring''' a__: Union[str, Any] = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( 'Tokenizer was not supporting padding necessary for zero-shot, attempting to use ' ' `pad_token=eos_token`') a__: Tuple = self.tokenizer.eos_token try: a__: Tuple = self.tokenizer( _A , add_special_tokens=_A , return_tensors=_A , padding=_A , truncation=_A , ) except Exception as e: if "too short" in str(_A): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. a__: Optional[int] = self.tokenizer( _A , add_special_tokens=_A , return_tensors=_A , padding=_A , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' if kwargs.get('multi_class' , _A) is not None: a__: Optional[int] = kwargs['multi_class'] logger.warning( 'The `multi_class` argument has been deprecated and renamed to `multi_label`. ' '`multi_class` will be removed in a future version of Transformers.') a__: int = {} if "candidate_labels" in kwargs: a__: Union[str, Any] = self._args_parser._parse_labels(kwargs['candidate_labels']) if "hypothesis_template" in kwargs: a__: int = kwargs['hypothesis_template'] a__: Dict = {} if "multi_label" in kwargs: a__: Union[str, Any] = kwargs['multi_label'] return preprocess_params, {}, postprocess_params def __call__( self , lowercase , *lowercase , **lowercase , ) -> Union[str, Any]: '''simple docstring''' if len(_A) == 0: pass elif len(_A) == 1 and "candidate_labels" not in kwargs: a__: Optional[int] = args[0] else: raise ValueError(f'Unable to understand extra arguments {args}') return super().__call__(_A , **_A) def lowerCamelCase_ ( self , lowercase , lowercase=None , lowercase="This example is {}.") -> Tuple: '''simple docstring''' a__: Union[str, Any] = self._args_parser(_A , _A , _A) for i, (candidate_label, sequence_pair) in enumerate(zip(_A , _A)): a__: Union[str, Any] = self._parse_and_tokenize([sequence_pair]) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(_A) - 1, **model_input, } def lowerCamelCase_ ( self , lowercase) -> Union[str, Any]: '''simple docstring''' a__: Dict = inputs['candidate_label'] a__: Optional[Any] = inputs['sequence'] a__: List[Any] = {k: inputs[k] for k in self.tokenizer.model_input_names} a__: Optional[Any] = self.model(**_A) a__: Optional[Any] = { 'candidate_label': candidate_label, 'sequence': sequence, 'is_last': inputs['is_last'], **outputs, } return model_outputs def lowerCamelCase_ ( self , lowercase , lowercase=False) -> Tuple: '''simple docstring''' a__: List[Any] = [outputs['candidate_label'] for outputs in model_outputs] a__: Optional[Any] = [outputs['sequence'] for outputs in model_outputs] a__: Optional[Any] = np.concatenate([output['logits'].numpy() for output in model_outputs]) a__: str = logits.shape[0] a__: Optional[int] = len(_A) a__: Tuple = N // n a__: Optional[int] = logits.reshape((num_sequences, n, -1)) if multi_label or len(_A) == 1: # softmax over the entailment vs. contradiction dim for each label independently a__: Union[str, Any] = self.entailment_id a__: Tuple = -1 if entailment_id == 0 else 0 a__: Dict = reshaped_outputs[..., [contradiction_id, entailment_id]] a__: Any = np.exp(_A) / np.exp(_A).sum(-1 , keepdims=_A) a__: List[str] = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels a__: Any = reshaped_outputs[..., self.entailment_id] a__: Union[str, Any] = np.exp(_A) / np.exp(_A).sum(-1 , keepdims=_A) a__: Union[str, Any] = list(reversed(scores[0].argsort())) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
290
import argparse import json from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( ) -> List[Any]: __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=a , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=a , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=a , help='where to store parsed gold_data_path file' , ) __A : Optional[int] = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: __A : List[Any] = json.load(a ) for dpr_record in tqdm(a ): __A : Dict = dpr_record['question'] __A : Any = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(a ) + '\n' ) if __name__ == "__main__": main()
280
0
import os import pytest from attr import dataclass SCREAMING_SNAKE_CASE__ : List[Any] = '''us-east-1''' # defaults region @dataclass class UpperCamelCase__ : '''simple docstring''' lowerCamelCase_ : str lowerCamelCase_ : Tuple = '''arn:aws:iam::558105141721:role/sagemaker_execution_role''' lowerCamelCase_ : str = { '''task_name''': '''mnli''', '''per_device_train_batch_size''': 1_6, '''per_device_eval_batch_size''': 1_6, '''do_train''': True, '''do_eval''': True, '''do_predict''': True, '''output_dir''': '''/opt/ml/model''', '''overwrite_output_dir''': True, '''max_steps''': 5_0_0, '''save_steps''': 5_5_0_0, } lowerCamelCase_ : Optional[int] = {**hyperparameters, '''max_steps''': 1_0_0_0} @property def _lowercase ( self ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def _lowercase ( self ) -> List[Any]: return F'''{self.framework}-transfromers-test''' @property def _lowercase ( self ) -> Any: return F'''./tests/sagemaker/scripts/{self.framework}''' @property def _lowercase ( self ) -> Dict: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="class" ) def A ( _SCREAMING_SNAKE_CASE ) -> Dict: lowerCamelCase : Optional[int] = SageMakerTestEnvironment(framework=request.cls.framework )
48
from heapq import heappop, heappush import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a , ) -> tuple[float | int, list[tuple[int, int]]]: __A , __A : int = grid.shape __A : Any = [-1, 1, 0, 0] __A : Optional[Any] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __A , __A : Optional[int] = [(0, source)], set() __A : Any = np.full((rows, cols) , np.inf ) __A : Any = 0 __A : Any = np.empty((rows, cols) , dtype=a ) __A : Optional[Any] = None while queue: ((__A) , (__A)) : List[str] = heappop(a ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __A : int = [] while (x, y) != source: path.append((x, y) ) __A , __A : Optional[int] = predecessors[x, y] path.append(a ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a ) ): __A , __A : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __A : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a , (dist + 1, (nx, ny)) ) __A : List[Any] = dist + 1 __A : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
280
0
# This is the module that test_patching.py uses to test patch_submodule() import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests lowercase = open # noqa: we just need to have a builtin inside this module to test it properly
178
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCamelCase_ = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
309
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _SCREAMING_SNAKE_CASE ( a , a ) -> Optional[Any]: __A : Any = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 10_24, 'hidden_size': 7_68, 'max_length': 5_12, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 10_24, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } __A : str = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __A : Optional[int] = BERTEncoder( attention_cell=predefined_args['attention_cell'] , num_layers=predefined_args['num_layers'] , units=predefined_args['units'] , hidden_size=predefined_args['hidden_size'] , max_length=predefined_args['max_length'] , num_heads=predefined_args['num_heads'] , scaled=predefined_args['scaled'] , dropout=predefined_args['dropout'] , output_attention=a , output_all_encodings=a , use_residual=predefined_args['use_residual'] , activation=predefined_args.get('activation' , 'gelu' ) , layer_norm_eps=predefined_args.get('layer_norm_eps' , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __A : Union[str, Any] = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab __A : Any = os.path.join(get_home_dir() , 'models' ) __A : List[Any] = _load_vocab(a , a , a , cls=a ) __A : Dict = nlp.model.BERTModel( a , len(a ) , units=predefined_args['units'] , embed_size=predefined_args['embed_size'] , embed_dropout=predefined_args['embed_dropout'] , word_embed=predefined_args['word_embed'] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args['token_type_vocab_size'] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) __A : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __A : Any = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(a ), } __A : int = BertConfig.from_dict(a ) __A : Union[str, Any] = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a , a ): __A : Tuple = hf_param.shape __A : str = to_torch(params[gluon_param] ) __A : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), F"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __A : str = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , 'word_embed.0.weight' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , 'encoder.position_weight' ) __A : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , 'encoder.layer_norm.beta' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , 'encoder.layer_norm.gamma' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __A : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __A : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __A : BertSelfAttention = layer.attention.self __A : Optional[Any] = check_and_map_params( self_attn.key.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.key.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.query.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __A : Optional[Any] = check_and_map_params( self_attn.query.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.value.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.value.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __A : BertSelfOutput = layer.attention.output __A : Tuple = check_and_map_params( self_output.dense.bias , F"""encoder.transformer_cells.{i}.proj.bias""" ) __A : int = check_and_map_params( self_output.dense.weight , F"""encoder.transformer_cells.{i}.proj.weight""" ) __A : List[Any] = check_and_map_params( self_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __A : str = check_and_map_params( self_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __A : BertIntermediate = layer.intermediate __A : int = check_and_map_params( intermediate.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __A : List[Any] = check_and_map_params( intermediate.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __A : BertOutput = layer.output __A : List[Any] = check_and_map_params( bert_output.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __A : Dict = check_and_map_params( bert_output.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __A : Optional[int] = check_and_map_params( bert_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __A : Dict = check_and_map_params( bert_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __A : Any = RobertaTokenizer.from_pretrained('roberta-base' ) __A : List[str] = tokenizer.encode_plus(a )['input_ids'] # Get gluon output __A : List[str] = mx.nd.array([input_ids] ) __A : Union[str, Any] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) __A : Optional[Any] = BertModel.from_pretrained(a ) hf_bort_model.eval() __A : Tuple = tokenizer.encode_plus(a , return_tensors='pt' ) __A : Any = hf_bort_model(**a )[0] __A : Union[str, Any] = output_gluon[0].asnumpy() __A : Tuple = output_hf[0].detach().numpy() __A : int = np.max(np.abs(hf_layer - gluon_layer ) ).item() __A : int = np.allclose(a , a , atol=1e-3 ) if success: print('✔️ Both model do output the same tensors' ) else: print('❌ Both model do **NOT** output the same tensors' ) print('Absolute difference is:' , a ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase : Dict = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
280
0
from __future__ import annotations import math def a ( A__ : Union[str, Any] , A__ : List[Any] , A__ : List[str] , A__ : Any , A__ : Dict ) -> int: """simple docstring""" if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(A__ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , A__ , A__ , A__ ) , minimax(depth + 1 , node_index * 2 + 1 , A__ , A__ , A__ ) , ) return min( minimax(depth + 1 , node_index * 2 , A__ , A__ , A__ ) , minimax(depth + 1 , node_index * 2 + 1 , A__ , A__ , A__ ) , ) def a ( ) -> None: """simple docstring""" _lowercase =[90, 23, 6, 33, 21, 65, 123, 34423] _lowercase =math.log(len(A__ ) , 2 ) print('Optimal value : ' , end='' ) print(minimax(0 , 0 , A__ , A__ , A__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
205
import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: __A : List[str] = x __A : str = y for step in range(a ): # noqa: B007 __A : Union[str, Any] = a * a - b * b + x __A : Optional[int] = 2 * a * b + y __A : List[str] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( a = 8_00 , a = 6_00 , a = -0.6 , a = 0 , a = 3.2 , a = 50 , a = True , ) -> Image.Image: __A : str = Image.new('RGB' , (image_width, image_height) ) __A : Dict = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates __A : Dict = figure_width / image_width * image_height __A : Union[str, Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width __A : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height __A : Union[str, Any] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __A : Optional[Any] = get_color_coded_rgb(a ) else: __A : Dict = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCAmelCase : str = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
280
0
'''simple docstring''' from collections.abc import Sequence def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[Any] = None ): '''simple docstring''' if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) UpperCAmelCase__ = nums[0] for i in range(1 , len(SCREAMING_SNAKE_CASE__ ) ): UpperCAmelCase__ = nums[i] UpperCAmelCase__ = max(SCREAMING_SNAKE_CASE__ , ans + num , SCREAMING_SNAKE_CASE__ ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user UpperCAmelCase_ = int(input('Enter number of elements : ').strip()) UpperCAmelCase_ = list(map(int, input('\nEnter the numbers : ').strip().split()))[:n] print(max_subsequence_sum(array))
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( a , nominal_annual_percentage_rate / 3_65 , number_of_years * 3_65 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Optional[Any] ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] ,model_result['ss'] ): UpperCAmelCase__ = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(_A ) def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=_A ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : List[Any] ): UpperCAmelCase__ = 'sgugger/tiny-distilbert-classification' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,only_pretrain_model=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(_A ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=_A ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ,[config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(_A ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ,[config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(_A ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ,[config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = 'patrickvonplaten/t5-tiny-random' UpperCAmelCase__ = AutoConfig.from_pretrained(_A ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ,configs=[config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('GPU' ) ) == 0 ,'Cannot do xla on CPU.' ) def __lowerCAmelCase ( self : Tuple ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=_A ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,use_xla=_A ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=_A ,save_to_csv=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,inference_time_csv_file=os.path.join(_A ,'inf_time.csv' ) ,inference_memory_csv_file=os.path.join(_A ,'inf_mem.csv' ) ,env_info_csv_file=os.path.join(_A ,'env.csv' ) ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) benchmark.run() self.assertTrue(Path(os.path.join(_A ,'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(_A ,'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(_A ,'env.csv' ) ).exists() ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(lowerCamelCase__ : List[str] ): self.assertTrue(hasattr(_A ,'sequential' ) ) self.assertTrue(hasattr(_A ,'cumulative' ) ) self.assertTrue(hasattr(_A ,'current' ) ) self.assertTrue(hasattr(_A ,'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=_A ,sequence_lengths=[8] ,batch_sizes=[1] ,log_filename=os.path.join(_A ,'log.txt' ) ,log_print=_A ,trace_memory_line_by_line=_A ,eager_mode=_A ,multi_process=_A ,) UpperCAmelCase__ = TensorFlowBenchmark(_A ) UpperCAmelCase__ = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(_A ,'log.txt' ) ).exists() )
98
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) _lowerCAmelCase : List[str] = 299_792_458 # Symbols _lowerCAmelCase : Optional[int] = symbols('''ct x y z''') def __snake_case ( _lowerCAmelCase : Tuple ) -> float: if velocity > c: raise ValueError("Speed must not exceed light speed 299,792,458 [m/s]!" ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError("Speed must be greater than or equal to 1!" ) return velocity / c def __snake_case ( _lowerCAmelCase : str ) -> float: return 1 / sqrt(1 - beta(_lowerCAmelCase ) ** 2 ) def __snake_case ( _lowerCAmelCase : str ) -> np.ndarray: return np.array( [ [gamma(_lowerCAmelCase ), -gamma(_lowerCAmelCase ) * beta(_lowerCAmelCase ), 0, 0], [-gamma(_lowerCAmelCase ) * beta(_lowerCAmelCase ), gamma(_lowerCAmelCase ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def __snake_case ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str = None ) -> np.ndarray: # Ensure event is not empty if event is None: A_ : str = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(_lowerCAmelCase ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: _lowerCAmelCase : str = transform(29_979_245) print('''Example of four vector: ''') print(F'''ct\' = {four_vector[0]}''') print(F'''x\' = {four_vector[1]}''') print(F'''y\' = {four_vector[2]}''') print(F'''z\' = {four_vector[3]}''') # Substitute symbols with numerical values _lowerCAmelCase : Union[str, Any] = {ct: c, x: 1, y: 1, z: 1} _lowerCAmelCase : Any = [four_vector[i].subs(sub_dict) for i in range(4)] print(F'''\n{numerical_vector}''')
300
def _SCREAMING_SNAKE_CASE ( a ) -> bool: return str(a ) == str(a )[::-1] def _SCREAMING_SNAKE_CASE ( a ) -> int: return int(a ) + int(str(a )[::-1] ) def _SCREAMING_SNAKE_CASE ( a = 1_00_00 ) -> int: __A : int = [] for num in range(1 , a ): __A : List[str] = 0 __A : List[Any] = num while iterations < 50: __A : str = sum_reverse(a ) iterations += 1 if is_palindrome(a ): break else: lychrel_nums.append(a ) return len(a ) if __name__ == "__main__": print(F"""{solution() = }""")
280
0
'''simple docstring''' import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("""0.8.3"""): raise Exception("""requires gluonnlp == 0.8.3""") if version.parse(mx.__version__) != version.parse("""1.5.0"""): raise Exception("""requires mxnet == 1.5.0""") logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) lowercase_ = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def lowerCamelCase ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[int] ) ->Optional[Any]: _SCREAMING_SNAKE_CASE = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 1024, 'hidden_size': 768, 'max_length': 512, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 1024, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } _SCREAMING_SNAKE_CASE = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py _SCREAMING_SNAKE_CASE = BERTEncoder( attention_cell=predefined_args["""attention_cell"""] , num_layers=predefined_args["""num_layers"""] , units=predefined_args["""units"""] , hidden_size=predefined_args["""hidden_size"""] , max_length=predefined_args["""max_length"""] , num_heads=predefined_args["""num_heads"""] , scaled=predefined_args["""scaled"""] , dropout=predefined_args["""dropout"""] , output_attention=__lowerCamelCase , output_all_encodings=__lowerCamelCase , use_residual=predefined_args["""use_residual"""] , activation=predefined_args.get("""activation""" , """gelu""" ) , layer_norm_eps=predefined_args.get("""layer_norm_eps""" , __lowerCamelCase ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later _SCREAMING_SNAKE_CASE = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab _SCREAMING_SNAKE_CASE = os.path.join(get_home_dir() , """models""" ) _SCREAMING_SNAKE_CASE = _load_vocab(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , cls=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = nlp.model.BERTModel( __lowerCamelCase , len(__lowerCamelCase ) , units=predefined_args["""units"""] , embed_size=predefined_args["""embed_size"""] , embed_dropout=predefined_args["""embed_dropout"""] , word_embed=predefined_args["""word_embed"""] , use_pooler=__lowerCamelCase , use_token_type_embed=__lowerCamelCase , token_type_vocab_size=predefined_args["""token_type_vocab_size"""] , use_classifier=__lowerCamelCase , use_decoder=__lowerCamelCase , ) original_bort.load_parameters(__lowerCamelCase , cast_dtype=__lowerCamelCase , ignore_extra=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = original_bort._collect_params_with_prefix() # Build our config 🤗 _SCREAMING_SNAKE_CASE = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(__lowerCamelCase ), } _SCREAMING_SNAKE_CASE = BertConfig.from_dict(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = BertForMaskedLM(__lowerCamelCase ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(__lowerCamelCase : Union[str, Any] ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(__lowerCamelCase : Optional[Any] , __lowerCamelCase : str ): _SCREAMING_SNAKE_CASE = hf_param.shape _SCREAMING_SNAKE_CASE = to_torch(params[gluon_param] ) _SCREAMING_SNAKE_CASE = gluon_param.shape assert ( shape_hf == shape_gluon ), F'The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers' return gluon_param _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , """word_embed.0.weight""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , """encoder.position_weight""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , """encoder.layer_norm.beta""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , """encoder.layer_norm.gamma""" ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) _SCREAMING_SNAKE_CASE = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): _SCREAMING_SNAKE_CASE = hf_bort_model.bert.encoder.layer[i] # self attention _SCREAMING_SNAKE_CASE = layer.attention.self _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.key.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_key.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.key.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_key.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.query.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_query.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.query.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_query.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.value.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_value.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.value.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_value.weight' ) # self attention output _SCREAMING_SNAKE_CASE = layer.attention.output _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.dense.bias , F'encoder.transformer_cells.{i}.proj.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.dense.weight , F'encoder.transformer_cells.{i}.proj.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.LayerNorm.bias , F'encoder.transformer_cells.{i}.layer_norm.beta' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.LayerNorm.weight , F'encoder.transformer_cells.{i}.layer_norm.gamma' ) # intermediate _SCREAMING_SNAKE_CASE = layer.intermediate _SCREAMING_SNAKE_CASE = check_and_map_params( intermediate.dense.bias , F'encoder.transformer_cells.{i}.ffn.ffn_1.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( intermediate.dense.weight , F'encoder.transformer_cells.{i}.ffn.ffn_1.weight' ) # output _SCREAMING_SNAKE_CASE = layer.output _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.dense.bias , F'encoder.transformer_cells.{i}.ffn.ffn_2.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.dense.weight , F'encoder.transformer_cells.{i}.ffn.ffn_2.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.LayerNorm.bias , F'encoder.transformer_cells.{i}.ffn.layer_norm.beta' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.LayerNorm.weight , F'encoder.transformer_cells.{i}.ffn.layer_norm.gamma' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models _SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained("""roberta-base""" ) _SCREAMING_SNAKE_CASE = tokenizer.encode_plus(__lowerCamelCase )['input_ids'] # Get gluon output _SCREAMING_SNAKE_CASE = mx.nd.array([input_ids] ) _SCREAMING_SNAKE_CASE = original_bort(inputs=__lowerCamelCase , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = BertModel.from_pretrained(__lowerCamelCase ) hf_bort_model.eval() _SCREAMING_SNAKE_CASE = tokenizer.encode_plus(__lowerCamelCase , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = hf_bort_model(**__lowerCamelCase )[0] _SCREAMING_SNAKE_CASE = output_gluon[0].asnumpy() _SCREAMING_SNAKE_CASE = output_hf[0].detach().numpy() _SCREAMING_SNAKE_CASE = np.max(np.abs(hf_layer - gluon_layer ) ).item() _SCREAMING_SNAKE_CASE = np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3 ) if success: print("""✔️ Both model do output the same tensors""" ) else: print("""❌ Both model do **NOT** output the same tensors""" ) print("""Absolute difference is:""" , __lowerCamelCase ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--bort_checkpoint_path""", default=None, type=str, required=True, help="""Path the official Bort params file.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase_ = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
58
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class _A: """simple docstring""" def __init__( self , _A = None ): if components is None: __A : int = [] __A : Tuple = list(_A ) def __len__( self ): return len(self.__components ) def __str__( self ): return "(" + ",".join(map(_A , self.__components ) ) + ")" def __add__( self , _A ): __A : Optional[int] = len(self ) if size == len(_A ): __A : Any = [self.__components[i] + other.component(_A ) for i in range(_A )] return Vector(_A ) else: raise Exception('must have the same size' ) def __sub__( self , _A ): __A : Tuple = len(self ) if size == len(_A ): __A : Union[str, Any] = [self.__components[i] - other.component(_A ) for i in range(_A )] return Vector(_A ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , (float, int) ): __A : str = [c * other for c in self.__components] return Vector(_A ) elif isinstance(_A , _A ) and len(self ) == len(_A ): __A : Union[str, Any] = len(self ) __A : Dict = [self.__components[i] * other.component(_A ) for i in range(_A )] return sum(_A ) else: # error case raise Exception('invalid operand!' ) def UpperCAmelCase_ ( self ): return Vector(self.__components ) def UpperCAmelCase_ ( self , _A ): if isinstance(_A , _A ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def UpperCAmelCase_ ( self , _A , _A ): assert -len(self.__components ) <= pos < len(self.__components ) __A : Optional[int] = value def UpperCAmelCase_ ( self ): if len(self.__components ) == 0: raise Exception('Vector is empty' ) __A : Optional[Any] = [c**2 for c in self.__components] return math.sqrt(sum(_A ) ) def UpperCAmelCase_ ( self , _A , _A = False ): __A : Optional[Any] = self * other __A : Optional[Any] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def _SCREAMING_SNAKE_CASE ( a ) -> Vector: assert isinstance(a , a ) return Vector([0] * dimension ) def _SCREAMING_SNAKE_CASE ( a , a ) -> Vector: assert isinstance(a , a ) and (isinstance(a , a )) __A : Optional[Any] = [0] * dimension __A : Tuple = 1 return Vector(a ) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: assert ( isinstance(a , a ) and isinstance(a , a ) and (isinstance(a , (int, float) )) ) return x * scalar + y def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: random.seed(a ) __A : str = [random.randint(a , a ) for _ in range(a )] return Vector(a ) class _A: """simple docstring""" def __init__( self , _A , _A , _A ): __A : Optional[Any] = matrix __A : Dict = w __A : Optional[int] = h def __str__( self ): __A : Tuple = '' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Optional[Any] = [] for i in range(self.__height ): __A : Optional[Any] = [ self.__matrix[i][j] + other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Tuple = [] for i in range(self.__height ): __A : str = [ self.__matrix[i][j] - other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , _A ): # matrix-vector if len(_A ) == self.__width: __A : List[Any] = zero_vector(self.__height ) for i in range(self.__height ): __A : List[str] = [ self.__matrix[i][j] * other.component(_A ) for j in range(self.__width ) ] ans.change_component(_A , sum(_A ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(_A , (int, float) ): # matrix-scalar __A : List[str] = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(_A , self.__width , self.__height ) return None def UpperCAmelCase_ ( self ): return self.__height def UpperCAmelCase_ ( self ): return self.__width def UpperCAmelCase_ ( self , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: __A : int = value else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) __A : List[str] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(_A ) ): __A : Optional[int] = minor[i][:y] + minor[i][y + 1 :] return Matrix(_A , self.__width - 1 , self.__height - 1 ).determinant() def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(_A , _A ) else: raise Exception('Indices out of bounds' ) def UpperCAmelCase_ ( self ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __A : List[str] = [ self.__matrix[0][y] * self.cofactor(0 , _A ) for y in range(self.__width ) ] return sum(_A ) def _SCREAMING_SNAKE_CASE ( a ) -> Matrix: __A : list[list[float]] = [[0] * n for _ in range(a )] return Matrix(a , a , a ) def _SCREAMING_SNAKE_CASE ( a , a , a , a ) -> Matrix: random.seed(a ) __A : list[list[float]] = [ [random.randint(a , a ) for _ in range(a )] for _ in range(a ) ] return Matrix(a , a , a )
280
0
from __future__ import annotations from dataclasses import dataclass @dataclass class _UpperCamelCase : '''simple docstring''' lowerCamelCase__ =42 lowerCamelCase__ =None lowerCamelCase__ =None def lowerCamelCase__ ( _a): # Validation def is_valid_tree(_a) -> bool: if node is None: return True if not isinstance(_a , _a): return False try: float(node.data) except (TypeError, ValueError): return False return is_valid_tree(node.left) and is_valid_tree(node.right) if not is_valid_tree(_a): raise ValueError( "Each node should be type of TreeNode and data should be float.") def is_binary_search_tree_recursive_check( _a , _a , _a) -> bool: if node is None: return True return ( left_bound < node.data < right_bound and is_binary_search_tree_recursive_check(node.left , _a , node.data) and is_binary_search_tree_recursive_check( node.right , node.data , _a) ) return is_binary_search_tree_recursive_check(_a , -float("inf") , float("inf")) if __name__ == "__main__": import doctest doctest.testmod()
76
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase : List[str] = '''▁''' UpperCAmelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Optional[int] = BertGenerationTokenizer UpperCamelCase : str = False UpperCamelCase : Tuple = True def UpperCAmelCase_ ( self ): super().setUp() __A : Tuple = BertGenerationTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : str = '<s>' __A : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def UpperCAmelCase_ ( self ): __A : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(_A ) , 1002 ) def UpperCAmelCase_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCAmelCase_ ( self ): __A : str = BertGenerationTokenizer(_A , keep_accents=_A ) __A : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_A , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [285, 46, 10, 170, 382] , ) __A : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __A : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __A : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase_ ( self ): return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def UpperCAmelCase_ ( self ): __A : List[Any] = 'Hello World!' __A : Optional[Any] = [18536, 2260, 101] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def UpperCAmelCase_ ( self ): __A : Dict = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) __A : int = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @require_torch @slow def UpperCAmelCase_ ( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence __A : Tuple = list(self.big_tokenizer.get_vocab().keys() )[:10] __A : List[Any] = ' '.join(_A ) __A : Union[str, Any] = self.big_tokenizer.encode_plus(_A , return_tensors='pt' , return_token_type_ids=_A ) __A : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=_A ) __A : int = BertGenerationConfig() __A : List[str] = BertGenerationEncoder(_A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def UpperCAmelCase_ ( self ): # fmt: off __A : str = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
280
0
from __future__ import annotations import math import random from typing import Any class A : def __init__(self : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" return self.head == self.tail def lowercase_ (self : List[Any] , __UpperCAmelCase : List[Any] ) -> Any: """simple docstring""" self.data.append(_A ) UpperCAmelCase__ = self.tail + 1 def lowercase_ (self : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.data[self.head] UpperCAmelCase__ = self.head + 1 return ret def lowercase_ (self : Dict ) -> Tuple: """simple docstring""" return self.tail - self.head def lowercase_ (self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" print(self.data ) print("**************" ) print(self.data[self.head : self.tail] ) class A : def __init__(self : Tuple , __UpperCAmelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = data UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = 1 def lowercase_ (self : List[str] ) -> str: """simple docstring""" return self.data def lowercase_ (self : Dict ) -> Tuple: """simple docstring""" return self.left def lowercase_ (self : Any ) -> Tuple: """simple docstring""" return self.right def lowercase_ (self : List[Any] ) -> str: """simple docstring""" return self.height def lowercase_ (self : Tuple , __UpperCAmelCase : Dict ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = data def lowercase_ (self : Any , __UpperCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = node def lowercase_ (self : List[str] , __UpperCAmelCase : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = node def lowercase_ (self : int , __UpperCAmelCase : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase__ = height def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if node is None: return 0 return node.get_height() def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' if a > b: return a return b def lowerCAmelCase_ ( __A ) -> MyNode: '''simple docstring''' print("left rotation node:", node.get_data() ) UpperCAmelCase__ = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(__A ) UpperCAmelCase__ = my_max(get_height(node.get_right() ), get_height(node.get_left() ) ) + 1 node.set_height(__A ) UpperCAmelCase__ = my_max(get_height(ret.get_right() ), get_height(ret.get_left() ) ) + 1 ret.set_height(__A ) return ret def lowerCAmelCase_ ( __A ) -> MyNode: '''simple docstring''' print("right rotation node:", node.get_data() ) UpperCAmelCase__ = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(__A ) UpperCAmelCase__ = my_max(get_height(node.get_right() ), get_height(node.get_left() ) ) + 1 node.set_height(__A ) UpperCAmelCase__ = my_max(get_height(ret.get_right() ), get_height(ret.get_left() ) ) + 1 ret.set_height(__A ) return ret def lowerCAmelCase_ ( __A ) -> MyNode: '''simple docstring''' UpperCAmelCase__ = node.get_left() assert left_child is not None node.set_left(left_rotation(__A ) ) return right_rotation(__A ) def lowerCAmelCase_ ( __A ) -> MyNode: '''simple docstring''' UpperCAmelCase__ = node.get_right() assert right_child is not None node.set_right(right_rotation(__A ) ) return left_rotation(__A ) def lowerCAmelCase_ ( __A, __A ) -> MyNode | None: '''simple docstring''' if node is None: return MyNode(__A ) if data < node.get_data(): node.set_left(insert_node(node.get_left(), __A ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected UpperCAmelCase__ = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child UpperCAmelCase__ = right_rotation(__A ) else: UpperCAmelCase__ = lr_rotation(__A ) else: node.set_right(insert_node(node.get_right(), __A ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: UpperCAmelCase__ = node.get_right() assert right_child is not None if data < right_child.get_data(): UpperCAmelCase__ = rl_rotation(__A ) else: UpperCAmelCase__ = left_rotation(__A ) UpperCAmelCase__ = my_max(get_height(node.get_right() ), get_height(node.get_left() ) ) + 1 node.set_height(__A ) return node def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' while True: UpperCAmelCase__ = root.get_right() if right_child is None: break UpperCAmelCase__ = right_child return root.get_data() def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' while True: UpperCAmelCase__ = root.get_left() if left_child is None: break UpperCAmelCase__ = left_child return root.get_data() def lowerCAmelCase_ ( __A, __A ) -> MyNode | None: '''simple docstring''' UpperCAmelCase__ = root.get_left() UpperCAmelCase__ = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: UpperCAmelCase__ = get_left_most(__A ) root.set_data(__A ) root.set_right(del_node(__A, __A ) ) elif left_child is not None: UpperCAmelCase__ = left_child elif right_child is not None: UpperCAmelCase__ = right_child else: return None elif root.get_data() > data: if left_child is None: print("No such data" ) return root else: root.set_left(del_node(__A, __A ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(__A, __A ) ) if get_height(__A ) - get_height(__A ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): UpperCAmelCase__ = left_rotation(__A ) else: UpperCAmelCase__ = rl_rotation(__A ) elif get_height(__A ) - get_height(__A ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): UpperCAmelCase__ = right_rotation(__A ) else: UpperCAmelCase__ = lr_rotation(__A ) UpperCAmelCase__ = my_max(get_height(root.get_right() ), get_height(root.get_left() ) ) + 1 root.set_height(__A ) return root class A : def __init__(self : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase__ = None def lowercase_ (self : str ) -> Union[str, Any]: """simple docstring""" return get_height(self.root ) def lowercase_ (self : Any , __UpperCAmelCase : Optional[Any] ) -> Tuple: """simple docstring""" print("insert:" + str(_A ) ) UpperCAmelCase__ = insert_node(self.root , _A ) def lowercase_ (self : List[Any] , __UpperCAmelCase : str ) -> Optional[Any]: """simple docstring""" print("delete:" + str(_A ) ) if self.root is None: print("Tree is empty!" ) return UpperCAmelCase__ = del_node(self.root , _A ) def __str__(self : List[str] , ) -> str: # a level traversale, gives a more intuitive look on the tree """simple docstring""" UpperCAmelCase__ = '' UpperCAmelCase__ = MyQueue() q.push(self.root ) UpperCAmelCase__ = self.get_height() if layer == 0: return output UpperCAmelCase__ = 0 while not q.is_empty(): UpperCAmelCase__ = q.pop() UpperCAmelCase__ = ' ' * int(math.pow(2 , layer - 1 ) ) output += space if node is None: output += "*" q.push(_A ) q.push(_A ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space UpperCAmelCase__ = cnt + 1 for i in range(1_0_0 ): if cnt == math.pow(2 , _A ) - 1: UpperCAmelCase__ = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def lowerCAmelCase_ ( ) -> None: '''simple docstring''' import doctest doctest.testmod() if __name__ == "__main__": _test() UpperCamelCase__ = AVLtree() UpperCamelCase__ = list(range(1_0)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
65
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _A: """simple docstring""" @staticmethod def UpperCAmelCase_ ( *_A , **_A ): pass def _SCREAMING_SNAKE_CASE ( a ) -> str: __A : str = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def _SCREAMING_SNAKE_CASE ( a ) -> Dict: __A : Dict = np.array(a ) __A : List[Any] = npimg.shape return {"hash": hashimage(a ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _A( unittest.TestCase ): """simple docstring""" UpperCamelCase : str = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase : int = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def UpperCAmelCase_ ( self , _A , _A , _A ): __A : Dict = MaskGenerationPipeline(model=_A , image_processor=_A ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase_ ( self , _A , _A ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def UpperCAmelCase_ ( self ): pass @slow @require_torch def UpperCAmelCase_ ( self ): __A : Union[str, Any] = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) __A : List[str] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=256 ) # Shortening by hashing __A : List[Any] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] , ) # fmt: on @require_torch @slow def UpperCAmelCase_ ( self ): __A : Optional[Any] = 'facebook/sam-vit-huge' __A : List[str] = pipeline('mask-generation' , model=_A ) __A : Tuple = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __A : List[str] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] , )
280
0
"""simple docstring""" from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10**-10 ) ->float: a__: Tuple = a while True: a__: List[Any] = Decimal(_SCREAMING_SNAKE_CASE ) - ( Decimal(eval(_SCREAMING_SNAKE_CASE ) ) / Decimal(eval(str(diff(_SCREAMING_SNAKE_CASE ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(_SCREAMING_SNAKE_CASE ) ) < precision: # noqa: S307 return float(_SCREAMING_SNAKE_CASE ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(f"The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}") # Find root of polynomial print(f"The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}") # Find Square Root of 5 print(f"The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}") # Exponential Roots print(f"The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}")
290
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): __A : List[Any] = tempfile.mkdtemp() # fmt: off __A : List[str] = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on __A : Union[str, Any] = dict(zip(_A , range(len(_A ) ) ) ) __A : Optional[int] = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] __A : int = {'unk_token': '<unk>'} __A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_A ) ) __A : List[Any] = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], 'image_std': [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } __A : Optional[int] = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def UpperCAmelCase_ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __A : Optional[int] = [Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase_ ( self ): __A : List[Any] = self.get_tokenizer() __A : str = self.get_rust_tokenizer() __A : List[str] = self.get_image_processor() __A : Optional[int] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_slow.save_pretrained(self.tmpdirname ) __A : int = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_fast.save_pretrained(self.tmpdirname ) __A : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _A ) self.assertIsInstance(processor_fast.tokenizer , _A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _A ) self.assertIsInstance(processor_fast.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : List[str] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) __A : Optional[int] = self.get_image_processor(do_normalize=_A ) __A : Any = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Optional[Any] = self.get_tokenizer() __A : Union[str, Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Union[str, Any] = self.prepare_image_inputs() __A : int = image_processor(_A , return_tensors='np' ) __A : str = processor(images=_A , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCAmelCase_ ( self ): __A : str = self.get_image_processor() __A : str = self.get_tokenizer() __A : Tuple = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : str = 'lower newer' __A : str = processor(text=_A , return_tensors='np' ) __A : List[str] = tokenizer(_A , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCAmelCase_ ( self ): __A : int = self.get_image_processor() __A : Optional[int] = self.get_tokenizer() __A : List[str] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Any = 'lower newer' __A : Optional[Any] = self.prepare_image_inputs() __A : List[Any] = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Any = 'google/owlvit-base-patch32' __A : int = OwlViTProcessor.from_pretrained(_A ) __A : Dict = ['cat', 'nasa badge'] __A : Optional[Any] = processor(text=_A ) __A : Optional[int] = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Tuple = 'google/owlvit-base-patch32' __A : Any = OwlViTProcessor.from_pretrained(_A ) __A : Dict = [['cat', 'nasa badge'], ['person']] __A : Dict = processor(text=_A ) __A : Optional[int] = 16 __A : Any = len(_A ) __A : Union[str, Any] = max([len(_A ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : List[Any] = 'google/owlvit-base-patch32' __A : str = OwlViTProcessor.from_pretrained(_A ) __A : Union[str, Any] = ['cat', 'nasa badge'] __A : Tuple = processor(text=_A ) __A : str = 16 __A : int = inputs['input_ids'] __A : List[Any] = [ [49406, 2368, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49406, 6841, 11301, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : List[str] = self.get_tokenizer() __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = processor(images=_A , query_images=_A ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Union[str, Any] = self.get_tokenizer() __A : str = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A : Any = processor.batch_decode(_A ) __A : Tuple = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A )
280
0
import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMSNForImageClassification, ViTMSNModel from transformers.models.vit_msn.modeling_vit_msn import VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__=13 , UpperCamelCase__=30 , UpperCamelCase__=2 , UpperCamelCase__=3 , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=32 , UpperCamelCase__=5 , UpperCamelCase__=4 , UpperCamelCase__=37 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=10 , UpperCamelCase__=0.02 , UpperCamelCase__=None , ) -> Tuple: lowerCamelCase : str = parent lowerCamelCase : int = batch_size lowerCamelCase : Any = image_size lowerCamelCase : Any = patch_size lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : Union[str, Any] = is_training lowerCamelCase : int = use_labels lowerCamelCase : Any = hidden_size lowerCamelCase : Tuple = num_hidden_layers lowerCamelCase : Any = num_attention_heads lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : int = hidden_act lowerCamelCase : Any = hidden_dropout_prob lowerCamelCase : List[str] = attention_probs_dropout_prob lowerCamelCase : Optional[Any] = type_sequence_label_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : Optional[Any] = scope # in ViT MSN, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase : List[Any] = (image_size // patch_size) ** 2 lowerCamelCase : Optional[Any] = num_patches + 1 def _lowercase ( self ) -> Tuple: lowerCamelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase : List[str] = None if self.use_labels: lowerCamelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : List[str] = self.get_config() return config, pixel_values, labels def _lowercase ( self ) -> Dict: return ViTMSNConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> str: lowerCamelCase : Optional[int] = ViTMSNModel(config=_A ) model.to(_A ) model.eval() lowerCamelCase : List[Any] = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: lowerCamelCase : str = self.type_sequence_label_size lowerCamelCase : Optional[Any] = ViTMSNForImageClassification(_A ) model.to(_A ) model.eval() lowerCamelCase : Dict = model(_A , labels=_A ) print("Pixel and labels shape: {pixel_values.shape}, {labels.shape}" ) print("Labels: {labels}" ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase : str = 1 lowerCamelCase : Any = ViTMSNForImageClassification(_A ) model.to(_A ) model.eval() lowerCamelCase : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase : int = model(_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self ) -> str: lowerCamelCase : int = self.prepare_config_and_inputs() lowerCamelCase : str = config_and_inputs lowerCamelCase : List[str] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ (snake_case__ , snake_case__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Optional[Any] = (ViTMSNModel, ViTMSNForImageClassification) if is_torch_available() else () lowerCamelCase_ : str = ( {'''feature-extraction''': ViTMSNModel, '''image-classification''': ViTMSNForImageClassification} if is_torch_available() else {} ) lowerCamelCase_ : Union[str, Any] = False lowerCamelCase_ : Tuple = False lowerCamelCase_ : str = False lowerCamelCase_ : List[str] = False def _lowercase ( self ) -> Optional[int]: lowerCamelCase : Dict = ViTMSNModelTester(self ) lowerCamelCase : Tuple = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=37 ) def _lowercase ( self ) -> Tuple: self.config_tester.run_common_tests() @unittest.skip(reason="ViTMSN does not use inputs_embeds" ) def _lowercase ( self ) -> List[str]: pass def _lowercase ( self ) -> List[str]: lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : int = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCamelCase : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear ) ) def _lowercase ( self ) -> Tuple: lowerCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : int = model_class(_A ) lowerCamelCase : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase : Any = [*signature.parameters.keys()] lowerCamelCase : Union[str, Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A ) def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def _lowercase ( self ) -> Any: lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) @slow def _lowercase ( self ) -> List[Any]: for model_name in VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase : str = ViTMSNModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def A ( ) -> Any: lowerCamelCase : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @cached_property def _lowercase ( self ) -> Tuple: return ViTImageProcessor.from_pretrained("facebook/vit-msn-small" ) if is_vision_available() else None @slow def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(2 ) lowerCamelCase : Union[str, Any] = ViTMSNForImageClassification.from_pretrained("facebook/vit-msn-small" ).to(_A ) lowerCamelCase : Dict = self.default_image_processor lowerCamelCase : Tuple = prepare_img() lowerCamelCase : List[Any] = image_processor(images=_A , return_tensors="pt" ).to(_A ) # forward pass with torch.no_grad(): lowerCamelCase : Union[str, Any] = model(**_A ) # verify the logits lowerCamelCase : int = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _A ) lowerCamelCase : str = torch.tensor([-0.0803, -0.4454, -0.2375] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1e-4 ) )
48
import math def _SCREAMING_SNAKE_CASE ( a ) -> list[int]: __A : List[str] = [] __A : Any = 2 __A : Union[str, Any] = int(math.sqrt(a ) ) # Size of every segment __A : Any = [True] * (end + 1) __A : List[Any] = [] while start <= end: if temp[start] is True: in_prime.append(a ) for i in range(start * start , end + 1 , a ): __A : Optional[int] = False start += 1 prime += in_prime __A : Any = end + 1 __A : Any = min(2 * end , a ) while low <= n: __A : List[Any] = [True] * (high - low + 1) for each in in_prime: __A : List[str] = math.floor(low / each ) * each if t < low: t += each for j in range(a , high + 1 , a ): __A : Optional[int] = False for j in range(len(a ) ): if temp[j] is True: prime.append(j + low ) __A : Optional[int] = high + 1 __A : Tuple = min(high + end , a ) return prime print(sieve(10**6))
280
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , a , a=7 , a=3 , a=30 , a=4_00 , a=True , a=None , a=0.9 , a=None , a=True , a=[0.5, 0.5, 0.5] , a=[0.5, 0.5, 0.5] , ) -> Optional[int]: snake_case_ = size if size is not None else {'shortest_edge': 30} snake_case_ = crop_size if crop_size is not None else {'height': 30, 'width': 30} snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = min_resolution snake_case_ = max_resolution snake_case_ = do_resize_and_center_crop snake_case_ = size snake_case_ = crop_pct snake_case_ = crop_size snake_case_ = do_normalize snake_case_ = image_mean snake_case_ = image_std def _UpperCamelCase ( self ) -> int: return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class UpperCamelCase_ ( snake_case__ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase = PoolFormerImageProcessor if is_vision_available() else None def _UpperCamelCase ( self ) -> Optional[Any]: snake_case_ = PoolFormerImageProcessingTester(self ) @property def _UpperCamelCase ( self ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self ) -> List[Any]: snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'do_resize_and_center_crop' ) ) self.assertTrue(hasattr(_A , 'size' ) ) self.assertTrue(hasattr(_A , 'crop_pct' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) def _UpperCamelCase ( self ) -> str: snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 30} ) self.assertEqual(image_processor.crop_size , {'height': 30, 'width': 30} ) snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def _UpperCamelCase ( self ) -> List[str]: pass def _UpperCamelCase ( self ) -> Union[str, Any]: # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def _UpperCamelCase ( self ) -> Tuple: # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def _UpperCamelCase ( self ) -> Optional[int]: # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
178
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : str = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
'''simple docstring''' UpperCamelCase_ = { '''a''': '''AAAAA''', '''b''': '''AAAAB''', '''c''': '''AAABA''', '''d''': '''AAABB''', '''e''': '''AABAA''', '''f''': '''AABAB''', '''g''': '''AABBA''', '''h''': '''AABBB''', '''i''': '''ABAAA''', '''j''': '''BBBAA''', '''k''': '''ABAAB''', '''l''': '''ABABA''', '''m''': '''ABABB''', '''n''': '''ABBAA''', '''o''': '''ABBAB''', '''p''': '''ABBBA''', '''q''': '''ABBBB''', '''r''': '''BAAAA''', '''s''': '''BAAAB''', '''t''': '''BAABA''', '''u''': '''BAABB''', '''v''': '''BBBAB''', '''w''': '''BABAA''', '''x''': '''BABAB''', '''y''': '''BABBA''', '''z''': '''BABBB''', ''' ''': ''' ''', } UpperCamelCase_ = {value: key for key, value in encode_dict.items()} def _UpperCAmelCase ( _lowerCamelCase : Tuple ) -> str: _lowerCAmelCase : Union[str, Any] = '' for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("""encode() accepts only letters of the alphabet and spaces""" ) return encoded def _UpperCAmelCase ( _lowerCamelCase : Optional[int] ) -> str: if set(_lowerCamelCase ) - {"A", "B", " "} != set(): raise Exception("""decode() accepts only \'A\', \'B\' and spaces""" ) _lowerCAmelCase : Any = '' for word in coded.split(): while len(_lowerCamelCase ) != 0: decoded += decode_dict[word[:5]] _lowerCAmelCase : Optional[int] = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
309
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def a ( A__ : List[Any] ) -> List[Any]: """simple docstring""" monkeypatch.setattr('datasets.utils.deprecation_utils._emitted_deprecation_warnings' , set() ) @pytest.fixture def a ( A__ : Dict ) -> List[str]: """simple docstring""" class __lowerCAmelCase : def __init__( self , lowerCAmelCase ) -> List[Any]: '''simple docstring''' _lowercase =metric_id class __lowerCAmelCase : _a = [MetricMock(snake_case__ ) for metric_id in ['''accuracy''', '''mse''', '''precision''', '''codeparrot/apps_metric''']] def A__ ( self ) -> List[str]: '''simple docstring''' return self._metrics monkeypatch.setattr('datasets.inspect.huggingface_hub' , HfhMock() ) @pytest.mark.parametrize( 'func, args' , [(load_metric, ('metrics/mse',)), (list_metrics, ()), (inspect_metric, ('metrics/mse', 'tmp_path'))] ) def a ( A__ : List[Any] , A__ : Optional[Any] , A__ : Union[str, Any] , A__ : Dict , A__ : Any ) -> int: """simple docstring""" if "tmp_path" in args: _lowercase =tuple(arg if arg != 'tmp_path' else tmp_path for arg in args ) with pytest.warns(A__ , match='https://huggingface.co/docs/evaluate' ): func(*A__ )
205
def _SCREAMING_SNAKE_CASE ( a , a = 0 ) -> list: __A : int = length or len(a ) __A : str = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __A , __A : Optional[int] = list_data[i + 1], list_data[i] __A : Union[str, Any] = True return list_data if not swapped else bubble_sort(a , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[Any] ): '''simple docstring''' if length <= 0 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError("""Length must be a positive integer.""" ) return [n * (2 * n - 1) for n in range(SCREAMING_SNAKE_CASE__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=1_0))
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a ) -> int: if not nums: return 0 __A : Optional[int] = nums[0] __A : str = 0 for num in nums[1:]: __A , __A : Tuple = ( max_excluding + num, max(a , a ), ) return max(a , a ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" import torch from diffusers import DiffusionPipeline class snake_case ( snake_case__ ): """simple docstring""" def __init__( self : Union[str, Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Any ): super().__init__() self.register_modules(unet=_A ,scheduler=_A ) def __call__( self : List[Any] ): UpperCAmelCase__ = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) ,) UpperCAmelCase__ = 1 UpperCAmelCase__ = self.unet(_A ,_A ).sample UpperCAmelCase__ = self.scheduler.step(_A ,_A ,_A ).prev_sample UpperCAmelCase__ = scheduler_output - scheduler_output + torch.ones_like(_A ) return result
98
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params _lowerCAmelCase : Dict = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ['''memory_attention''', '''encoder_attn'''], ['''attention''', '''attn'''], ['''/''', '''.'''], ['''.LayerNorm.gamma''', '''_layer_norm.weight'''], ['''.LayerNorm.beta''', '''_layer_norm.bias'''], ['''r.layer_''', '''r.layers.'''], ['''output_proj''', '''out_proj'''], ['''ffn.dense_1.''', '''fc2.'''], ['''ffn.dense.''', '''fc1.'''], ['''ffn_layer_norm''', '''final_layer_norm'''], ['''kernel''', '''weight'''], ['''encoder_layer_norm.''', '''encoder.layer_norm.'''], ['''decoder_layer_norm.''', '''decoder.layer_norm.'''], ['''embeddings.weights''', '''shared.weight'''], ] def __snake_case ( _lowerCAmelCase : Optional[Any] ) -> Dict: for pegasus_name, hf_name in PATTERNS: A_ : List[Any] = k.replace(_lowerCAmelCase , _lowerCAmelCase ) return k def __snake_case ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ) -> PegasusForConditionalGeneration: A_ : str = DEFAULTS.copy() cfg_kwargs.update(_lowerCAmelCase ) A_ : Optional[Any] = PegasusConfig(**_lowerCAmelCase ) A_ : int = PegasusForConditionalGeneration(_lowerCAmelCase ) A_ : List[str] = torch_model.model.state_dict() A_ : int = {} for k, v in tf_weights.items(): A_ : Optional[int] = rename_state_dict_key(_lowerCAmelCase ) if new_k not in sd: raise ValueError(f"could not find new key {new_k} in state dict. (converted from {k})" ) if "dense" in k or "proj" in new_k: A_ : Optional[int] = v.T A_ : Optional[int] = torch.tensor(_lowerCAmelCase , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, f"{new_k}, {k}, {v.shape}, {sd[new_k].shape}" # make sure embedding.padding_idx is respected A_ : Optional[Any] = torch.zeros_like(mapping["shared.weight"][cfg.pad_token_id + 1] ) A_ : Optional[int] = mapping['shared.weight'] A_ : Optional[Any] = mapping['shared.weight'] A_ : List[str] = {k: torch.zeros_like(_lowerCAmelCase ) for k, v in sd.items() if k.endswith("bias" ) and k not in mapping} mapping.update(**_lowerCAmelCase ) A_ : Any = torch_model.model.load_state_dict(_lowerCAmelCase , strict=_lowerCAmelCase ) A_ : Any = [ k for k in missing if k not in ['encoder.embed_positions.weight', 'decoder.embed_positions.weight'] ] assert unexpected_missing == [], f"no matches found for the following torch keys {unexpected_missing}" assert extra == [], f"no matches found for the following tf keys {extra}" return torch_model def __snake_case ( _lowerCAmelCase : Any="./ckpt/aeslc/model.ckpt-32000" ) -> Dict: A_ : List[Any] = tf.train.list_variables(_lowerCAmelCase ) A_ : Union[str, Any] = {} A_ : Optional[Any] = ['Adafactor', 'global_step'] for name, shape in tqdm(_lowerCAmelCase , desc="converting tf checkpoint to dict" ): A_ : List[Any] = any(pat in name for pat in ignore_name ) if skip_key: continue A_ : Optional[int] = tf.train.load_variable(_lowerCAmelCase , _lowerCAmelCase ) A_ : Dict = array return tf_weights def __snake_case ( _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ) -> str: # save tokenizer first A_ : List[str] = Path(_lowerCAmelCase ).parent.name A_ : Optional[Any] = task_specific_params[f"summarization_{dataset}"]['max_position_embeddings'] A_ : str = PegasusTokenizer.from_pretrained("sshleifer/pegasus" , model_max_length=_lowerCAmelCase ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(_lowerCAmelCase ) # convert model A_ : Optional[Any] = get_tf_weights_as_numpy(_lowerCAmelCase ) A_ : Any = task_specific_params[f"summarization_{dataset}"] if dataset == "large": A_ : Tuple = task_specific_params A_ : Tuple = convert_pegasus(_lowerCAmelCase , _lowerCAmelCase ) torch_model.save_pretrained(_lowerCAmelCase ) A_ : Dict = torch_model.state_dict() sd.pop("model.decoder.embed_positions.weight" ) sd.pop("model.encoder.embed_positions.weight" ) torch.save(_lowerCAmelCase , Path(_lowerCAmelCase ) / "pytorch_model.bin" ) if __name__ == "__main__": _lowerCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument('''tf_ckpt_path''', type=str, help='''passed to tf.train.list_variables''') parser.add_argument('''save_dir''', default=None, type=str, help='''Path to the output PyTorch model.''') _lowerCAmelCase : Optional[Any] = parser.parse_args() if args.save_dir is None: _lowerCAmelCase : str = Path(args.tf_ckpt_path).parent.name _lowerCAmelCase : Optional[int] = os.path.join('''pegasus''', dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
300
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' def lowerCamelCase ( __lowerCamelCase : List[Any] ) ->list[int]: _SCREAMING_SNAKE_CASE = [0 for i in range(len(__lowerCamelCase ) )] # initialize interval's left pointer and right pointer _SCREAMING_SNAKE_CASE = 0, 0 for i in range(1 , len(__lowerCamelCase ) ): # case when current index is inside the interval if i <= right_pointer: _SCREAMING_SNAKE_CASE = min(right_pointer - i + 1 , z_result[i - left_pointer] ) _SCREAMING_SNAKE_CASE = min_edge while go_next(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: _SCREAMING_SNAKE_CASE = i, i + z_result[i] - 1 return z_result def lowerCamelCase ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] ) ->bool: return i + z_result[i] < len(__lowerCamelCase ) and s[z_result[i]] == s[i + z_result[i]] def lowerCamelCase ( __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] ) ->int: _SCREAMING_SNAKE_CASE = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string _SCREAMING_SNAKE_CASE = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(__lowerCamelCase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
58
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> None: __A : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a ) == len(a ), F"""{len(a )} != {len(a )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) UpperCAmelCase : List[Any] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } UpperCAmelCase : Optional[int] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _SCREAMING_SNAKE_CASE ( a , a ) -> Dict: try: __A : int = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a ) ) def _SCREAMING_SNAKE_CASE ( a , a ) -> List[int]: if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _SCREAMING_SNAKE_CASE ( a , a = "student" , a = None , a = None , a=False , a=None , a=None , **a , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __A : List[str] = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(a , a ): AutoTokenizer.from_pretrained(a ).save_pretrained(a ) # purely for convenience __A : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained(a ).eval() else: assert isinstance(a , a ), F"""teacher must be a model or string got type {type(a )}""" __A : int = teacher.config.to_diff_dict() try: __A , __A : List[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A : str = teacher_e if d is None: __A : List[Any] = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): __A , __A : List[Any] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A : Optional[int] = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A : int = teacher_e if d is None: __A : Optional[Any] = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a ) # Copy weights __A : Dict = teacher.config_class(**a ) __A : int = AutoModelForSeqaSeqLM.from_config(a ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A : Any = student.load_state_dict(teacher.state_dict() , strict=a ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A : Optional[int] = list(range(a ) ), list(range(a ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) if d_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) try: if hasattr( a , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a ) copy_layers(teacher.decoder.block , student.decoder.block , a ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __A : Optional[int] = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(a ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
280
0
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor a_ = logging.get_logger(__name__) class _UpperCamelCase ( snake_case__ ): '''simple docstring''' def __init__( self : Optional[Any] , *a : Any , **a : Any ) -> Dict: """simple docstring""" warnings.warn( "The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use CLIPImageProcessor instead." , _A , ) super().__init__(*_A , **_A )
76
def _SCREAMING_SNAKE_CASE ( a , a ) -> list[int]: __A : Optional[int] = int(a ) # Initialize Result __A : Optional[int] = [] # Traverse through all denomination for denomination in reversed(a ): # Find denominations while int(a ) >= int(a ): total_value -= int(a ) answer.append(a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase : List[str] = [] UpperCAmelCase : Optional[int] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): UpperCAmelCase : List[Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) UpperCAmelCase : int = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCAmelCase : Tuple = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(F"""Following is minimal change for {value}: """) UpperCAmelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
280
0
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { '''deepmind/language-perceiver''': '''https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json''', # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class A ( snake_case__ ): __UpperCAmelCase : Optional[int] = '''perceiver''' def __init__(self : Any , __UpperCAmelCase : Optional[Any]=2_5_6 , __UpperCAmelCase : List[str]=1_2_8_0 , __UpperCAmelCase : Dict=7_6_8 , __UpperCAmelCase : Union[str, Any]=1 , __UpperCAmelCase : int=2_6 , __UpperCAmelCase : List[Any]=8 , __UpperCAmelCase : Any=8 , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Any=None , __UpperCAmelCase : List[str]="kv" , __UpperCAmelCase : Optional[Any]=1 , __UpperCAmelCase : int=1 , __UpperCAmelCase : int="gelu" , __UpperCAmelCase : Optional[Any]=0.1 , __UpperCAmelCase : Tuple=0.02 , __UpperCAmelCase : List[Any]=1E-12 , __UpperCAmelCase : Any=True , __UpperCAmelCase : List[str]=2_6_2 , __UpperCAmelCase : Dict=2_0_4_8 , __UpperCAmelCase : Optional[Any]=5_6 , __UpperCAmelCase : Optional[int]=[3_6_8, 4_9_6] , __UpperCAmelCase : Any=1_6 , __UpperCAmelCase : List[Any]=1_9_2_0 , __UpperCAmelCase : Tuple=1_6 , __UpperCAmelCase : List[str]=[1, 1_6, 2_2_4, 2_2_4] , **__UpperCAmelCase : Optional[Any] , ) -> Dict: """simple docstring""" super().__init__(**_A ) UpperCAmelCase__ = num_latents UpperCAmelCase__ = d_latents UpperCAmelCase__ = d_model UpperCAmelCase__ = num_blocks UpperCAmelCase__ = num_self_attends_per_block UpperCAmelCase__ = num_self_attention_heads UpperCAmelCase__ = num_cross_attention_heads UpperCAmelCase__ = qk_channels UpperCAmelCase__ = v_channels UpperCAmelCase__ = cross_attention_shape_for_attention UpperCAmelCase__ = self_attention_widening_factor UpperCAmelCase__ = cross_attention_widening_factor UpperCAmelCase__ = hidden_act UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = use_query_residual # masked language modeling attributes UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings # image classification attributes UpperCAmelCase__ = image_size # flow attributes UpperCAmelCase__ = train_size # multimodal autoencoding attributes UpperCAmelCase__ = num_frames UpperCAmelCase__ = audio_samples_per_frame UpperCAmelCase__ = samples_per_patch UpperCAmelCase__ = output_shape class A ( snake_case__ ): @property def lowercase_ (self : Any ) -> Optional[int]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase__ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: UpperCAmelCase__ = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ("inputs", dynamic_axis), ("attention_mask", dynamic_axis), ] ) @property def lowercase_ (self : str ) -> List[str]: """simple docstring""" return 1E-4 def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple = -1 , __UpperCAmelCase : int = -1 , __UpperCAmelCase : Any = -1 , __UpperCAmelCase : Dict = False , __UpperCAmelCase : Dict = None , __UpperCAmelCase : str = 3 , __UpperCAmelCase : Any = 4_0 , __UpperCAmelCase : List[Any] = 4_0 , ) -> Dict: """simple docstring""" if isinstance(_A , _A ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase__ = compute_effective_axis_dimension( _A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase__ = preprocessor.num_special_tokens_to_add(_A ) UpperCAmelCase__ = compute_effective_axis_dimension( _A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_A ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase__ = [' '.join(["a"] ) * seq_length] * batch_size UpperCAmelCase__ = dict(preprocessor(_A , return_tensors=_A ) ) UpperCAmelCase__ = inputs.pop("input_ids" ) return inputs elif isinstance(_A , _A ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase__ = compute_effective_axis_dimension(_A , fixed_dimension=OnnxConfig.default_fixed_batch ) UpperCAmelCase__ = self._generate_dummy_images(_A , _A , _A , _A ) UpperCAmelCase__ = dict(preprocessor(images=_A , return_tensors=_A ) ) UpperCAmelCase__ = inputs.pop("pixel_values" ) return inputs else: raise ValueError( "Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor." )
65
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _A( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __A : List[Any] = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __A : Union[str, Any] = parent __A : Optional[int] = batch_size __A : int = num_channels __A : int = min_resolution __A : Any = max_resolution __A : List[Any] = do_resize __A : List[Any] = size __A : Union[str, Any] = do_normalize __A : Optional[int] = image_mean __A : Optional[int] = image_std __A : int = do_rescale __A : str = rescale_factor __A : Tuple = do_pad def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase_ ( self , _A , _A=False ): if not batched: __A : List[str] = image_inputs[0] if isinstance(_A , Image.Image ): __A , __A : int = image.size else: __A , __A : Any = image.shape[1], image.shape[2] if w < h: __A : List[Any] = int(self.size['shortest_edge'] * h / w ) __A : List[Any] = self.size['shortest_edge'] elif w > h: __A : Union[str, Any] = self.size['shortest_edge'] __A : str = int(self.size['shortest_edge'] * w / h ) else: __A : Dict = self.size['shortest_edge'] __A : str = self.size['shortest_edge'] else: __A : int = [] for image in image_inputs: __A , __A : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __A : List[str] = max(_A , key=lambda _A : item[0] )[0] __A : str = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : List[str] = YolosImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): __A : Dict = YolosImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): __A : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) def UpperCAmelCase_ ( self ): __A : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , _A ) __A : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _A ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing __A : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input __A : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A , __A : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) __A : str = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input __A : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : List[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Tuple = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input __A : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Union[str, Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Optional[int] = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processings __A : Tuple = self.image_processing_class(**self.image_processor_dict ) __A : Any = self.image_processing_class(do_resize=_A , do_normalize=_A , do_rescale=_A ) # create random PyTorch tensors __A : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __A : Optional[int] = image_processing_a.pad(_A , return_tensors='pt' ) __A : Optional[int] = image_processing_a(_A , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1e-4 ) ) @slow def UpperCAmelCase_ ( self ): # prepare image and target __A : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __A : Optional[Any] = json.loads(f.read() ) __A : Optional[Any] = {'image_id': 39769, 'annotations': target} # encode them __A : str = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) __A : List[Any] = image_processing(images=_A , annotations=_A , return_tensors='pt' ) # verify pixel values __A : List[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : List[Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Any = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify orig_size __A : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : str = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) ) @slow def UpperCAmelCase_ ( self ): # prepare image, target and masks_path __A : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __A : Tuple = json.loads(f.read() ) __A : Any = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __A : List[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __A : Any = YolosImageProcessor(format='coco_panoptic' ) __A : List[Any] = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='pt' ) # verify pixel values __A : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Union[str, Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : List[str] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify masks __A : Tuple = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _A ) # verify orig_size __A : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : int = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) )
280
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase__ = logging.get_logger(__name__) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: int = b.T a__: List[Any] = np.sum(np.square(_SCREAMING_SNAKE_CASE ) , axis=1 ) a__: int = np.sum(np.square(_SCREAMING_SNAKE_CASE ) , axis=0 ) a__: Any = np.matmul(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Optional[Any] = aa[:, None] - 2 * ab + ba[None, :] return d def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Any: a__: Tuple = x.reshape(-1 , 3 ) a__: Any = squared_euclidean_distance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return np.argmin(_SCREAMING_SNAKE_CASE , axis=1 ) class __snake_case ( snake_case__ ): a__ = ['''pixel_values'''] def __init__( self , lowercase = None , lowercase = True , lowercase = None , lowercase = PILImageResampling.BILINEAR , lowercase = True , lowercase = True , **lowercase , ) -> Optional[Any]: '''simple docstring''' super().__init__(**_A) a__: Optional[Any] = size if size is not None else {'height': 2_56, 'width': 2_56} a__: Any = get_size_dict(_A) a__: List[str] = np.array(_A) if clusters is not None else None a__: int = do_resize a__: Union[str, Any] = size a__: Dict = resample a__: List[Any] = do_normalize a__: str = do_color_quantize def lowerCamelCase_ ( self , lowercase , lowercase , lowercase = PILImageResampling.BILINEAR , lowercase = None , **lowercase , ) -> int: '''simple docstring''' a__: Optional[Any] = get_size_dict(_A) if "height" not in size or "width" not in size: raise ValueError(f'Size dictionary must contain both height and width keys. Got {size.keys()}') return resize( _A , size=(size['height'], size['width']) , resample=_A , data_format=_A , **_A) def lowerCamelCase_ ( self , lowercase , lowercase = None , ) -> Dict: '''simple docstring''' a__: Dict = rescale(image=_A , scale=1 / 1_27.5 , data_format=_A) a__: str = image - 1 return image def lowerCamelCase_ ( self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = ChannelDimension.FIRST , **lowercase , ) -> Union[str, Any]: '''simple docstring''' a__: Any = do_resize if do_resize is not None else self.do_resize a__: Dict = size if size is not None else self.size a__: Union[str, Any] = get_size_dict(_A) a__: Optional[Any] = resample if resample is not None else self.resample a__: Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize a__: Optional[int] = do_color_quantize if do_color_quantize is not None else self.do_color_quantize a__: List[str] = clusters if clusters is not None else self.clusters a__: str = np.array(_A) a__: int = make_list_of_images(_A) if not valid_images(_A): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.') if do_color_quantize and clusters is None: raise ValueError('Clusters must be specified if do_color_quantize is True.') # All transformations expect numpy arrays. a__: str = [to_numpy_array(_A) for image in images] if do_resize: a__: int = [self.resize(image=_A , size=_A , resample=_A) for image in images] if do_normalize: a__: List[Any] = [self.normalize(image=_A) for image in images] if do_color_quantize: a__: List[Any] = [to_channel_dimension_format(_A , ChannelDimension.LAST) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) a__: List[Any] = np.array(_A) a__: str = color_quantize(_A , _A).reshape(images.shape[:-1]) # flatten to (batch_size, height*width) a__: Dict = images.shape[0] a__: List[Any] = images.reshape(_A , -1) # We need to convert back to a list of images to keep consistent behaviour across processors. a__: Union[str, Any] = list(_A) else: a__: Any = [to_channel_dimension_format(_A , _A) for image in images] a__: List[Any] = {'input_ids': images} return BatchFeature(data=_A , tensor_type=_A)
290
import argparse import json from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( ) -> List[Any]: __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=a , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=a , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=a , help='where to store parsed gold_data_path file' , ) __A : Optional[int] = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: __A : List[Any] = json.load(a ) for dpr_record in tqdm(a ): __A : Dict = dpr_record['question'] __A : Any = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(a ) + '\n' ) if __name__ == "__main__": main()
280
0
import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef SCREAMING_SNAKE_CASE__ : Optional[Any] = ( '''This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: warnings.warn(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) requires_backends(_SCREAMING_SNAKE_CASE ,"sklearn" ) return (preds == labels).mean() def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: warnings.warn(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) requires_backends(_SCREAMING_SNAKE_CASE ,"sklearn" ) lowerCamelCase : Any = simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) lowerCamelCase : Dict = fa_score(y_true=_SCREAMING_SNAKE_CASE ,y_pred=_SCREAMING_SNAKE_CASE ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> str: warnings.warn(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) requires_backends(_SCREAMING_SNAKE_CASE ,"sklearn" ) lowerCamelCase : int = pearsonr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )[0] lowerCamelCase : List[str] = spearmanr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: warnings.warn(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) requires_backends(_SCREAMING_SNAKE_CASE ,"sklearn" ) assert len(_SCREAMING_SNAKE_CASE ) == len(_SCREAMING_SNAKE_CASE ), f'''Predictions and labels have mismatched lengths {len(_SCREAMING_SNAKE_CASE )} and {len(_SCREAMING_SNAKE_CASE )}''' if task_name == "cola": return {"mcc": matthews_corrcoef(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "sst-2": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "mrpc": return acc_and_fa(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) elif task_name == "sts-b": return pearson_and_spearman(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) elif task_name == "qqp": return acc_and_fa(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "qnli": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "rte": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "wnli": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} elif task_name == "hans": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} else: raise KeyError(_SCREAMING_SNAKE_CASE ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> List[Any]: warnings.warn(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) requires_backends(_SCREAMING_SNAKE_CASE ,"sklearn" ) if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError(f'''Predictions and labels have mismatched lengths {len(_SCREAMING_SNAKE_CASE )} and {len(_SCREAMING_SNAKE_CASE )}''' ) if task_name == "xnli": return {"acc": simple_accuracy(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE )} else: raise KeyError(_SCREAMING_SNAKE_CASE )
48
from heapq import heappop, heappush import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a , ) -> tuple[float | int, list[tuple[int, int]]]: __A , __A : int = grid.shape __A : Any = [-1, 1, 0, 0] __A : Optional[Any] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __A , __A : Optional[int] = [(0, source)], set() __A : Any = np.full((rows, cols) , np.inf ) __A : Any = 0 __A : Any = np.empty((rows, cols) , dtype=a ) __A : Optional[Any] = None while queue: ((__A) , (__A)) : List[str] = heappop(a ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __A : int = [] while (x, y) != source: path.append((x, y) ) __A , __A : Optional[int] = predecessors[x, y] path.append(a ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a ) ): __A , __A : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __A : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a , (dist + 1, (nx, ny)) ) __A : List[Any] = dist + 1 __A : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
280
0
import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def _UpperCamelCase ( self ) -> str: snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_A ) snake_case_ = -1 snake_case_ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_A ) snake_case_ = model.generate(_A , max_new_tokens=10 , do_sample=_A ) snake_case_ = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: snake_case_ = TextStreamer(_A ) model.generate(_A , max_new_tokens=10 , do_sample=_A , streamer=_A ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer snake_case_ = cs.out[:-1] self.assertEqual(_A , _A ) def _UpperCamelCase ( self ) -> str: snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_A ) snake_case_ = -1 snake_case_ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_A ) snake_case_ = model.generate(_A , max_new_tokens=10 , do_sample=_A ) snake_case_ = tokenizer.decode(greedy_ids[0] ) snake_case_ = TextIteratorStreamer(_A ) snake_case_ = {'input_ids': input_ids, 'max_new_tokens': 10, 'do_sample': False, 'streamer': streamer} snake_case_ = Thread(target=model.generate , kwargs=_A ) thread.start() snake_case_ = '' for new_text in streamer: streamer_text += new_text self.assertEqual(_A , _A ) def _UpperCamelCase ( self ) -> str: snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_A ) snake_case_ = -1 snake_case_ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_A ) snake_case_ = model.generate(_A , max_new_tokens=10 , do_sample=_A ) snake_case_ = greedy_ids[:, input_ids.shape[1] :] snake_case_ = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: snake_case_ = TextStreamer(_A , skip_prompt=_A ) model.generate(_A , max_new_tokens=10 , do_sample=_A , streamer=_A ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer snake_case_ = cs.out[:-1] self.assertEqual(_A , _A ) def _UpperCamelCase ( self ) -> List[Any]: # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them snake_case_ = AutoTokenizer.from_pretrained('distilgpt2' ) snake_case_ = AutoModelForCausalLM.from_pretrained('distilgpt2' ).to(_A ) snake_case_ = -1 snake_case_ = torch.ones((1, 5) , device=_A ).long() * model.config.bos_token_id with CaptureStdout() as cs: snake_case_ = TextStreamer(_A , skip_special_tokens=_A ) model.generate(_A , max_new_tokens=1 , do_sample=_A , streamer=_A ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token snake_case_ = cs.out[:-1] # Remove the final "\n" snake_case_ = tokenizer(_A , return_tensors='pt' ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def _UpperCamelCase ( self ) -> List[str]: snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_A ) snake_case_ = -1 snake_case_ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_A ) snake_case_ = TextIteratorStreamer(_A , timeout=0.001 ) snake_case_ = {'input_ids': input_ids, 'max_new_tokens': 10, 'do_sample': False, 'streamer': streamer} snake_case_ = Thread(target=model.generate , kwargs=_A ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(_A ): snake_case_ = '' for new_text in streamer: streamer_text += new_text
178
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
0
'''simple docstring''' import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class a_ (unittest.TestCase ): def __UpperCamelCase ( self ): _lowerCAmelCase : str = inspect.getfile(accelerate.test_utils ) _lowerCAmelCase : Optional[int] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) _lowerCAmelCase : List[str] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) _lowerCAmelCase : List[Any] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def __UpperCamelCase ( self ): print(f'Found {torch.cuda.device_count()} devices.' ) _lowerCAmelCase : Tuple = ['torchrun', f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_A , env=os.environ.copy() ) @require_multi_gpu def __UpperCamelCase ( self ): print(f'Found {torch.cuda.device_count()} devices.' ) _lowerCAmelCase : Optional[Any] = ['torchrun', f'--nproc_per_node={torch.cuda.device_count()}', self.operation_file_path] print(f'Command: {cmd}' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_A , env=os.environ.copy() ) @require_multi_gpu def __UpperCamelCase ( self ): _lowerCAmelCase : Optional[Any] = ['torchrun', f'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_A , env=os.environ.copy() ) @require_multi_gpu def __UpperCamelCase ( self ): print(f'Found {torch.cuda.device_count()} devices, using 2 devices only' ) _lowerCAmelCase : List[Any] = ['torchrun', f'--nproc_per_node={torch.cuda.device_count()}', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(_A , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase_ = Accelerator() UpperCamelCase_ = (accelerator.state.process_index + 2, 10) UpperCamelCase_ = torch.randint(0, 10, shape).to(accelerator.device) UpperCamelCase_ = '''''' UpperCamelCase_ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." UpperCamelCase_ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." UpperCamelCase_ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
309
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _SCREAMING_SNAKE_CASE ( a , a ) -> Optional[Any]: __A : Any = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 10_24, 'hidden_size': 7_68, 'max_length': 5_12, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 10_24, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } __A : str = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __A : Optional[int] = BERTEncoder( attention_cell=predefined_args['attention_cell'] , num_layers=predefined_args['num_layers'] , units=predefined_args['units'] , hidden_size=predefined_args['hidden_size'] , max_length=predefined_args['max_length'] , num_heads=predefined_args['num_heads'] , scaled=predefined_args['scaled'] , dropout=predefined_args['dropout'] , output_attention=a , output_all_encodings=a , use_residual=predefined_args['use_residual'] , activation=predefined_args.get('activation' , 'gelu' ) , layer_norm_eps=predefined_args.get('layer_norm_eps' , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __A : Union[str, Any] = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab __A : Any = os.path.join(get_home_dir() , 'models' ) __A : List[Any] = _load_vocab(a , a , a , cls=a ) __A : Dict = nlp.model.BERTModel( a , len(a ) , units=predefined_args['units'] , embed_size=predefined_args['embed_size'] , embed_dropout=predefined_args['embed_dropout'] , word_embed=predefined_args['word_embed'] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args['token_type_vocab_size'] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) __A : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __A : Any = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(a ), } __A : int = BertConfig.from_dict(a ) __A : Union[str, Any] = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a , a ): __A : Tuple = hf_param.shape __A : str = to_torch(params[gluon_param] ) __A : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), F"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __A : str = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , 'word_embed.0.weight' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , 'encoder.position_weight' ) __A : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , 'encoder.layer_norm.beta' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , 'encoder.layer_norm.gamma' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __A : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __A : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __A : BertSelfAttention = layer.attention.self __A : Optional[Any] = check_and_map_params( self_attn.key.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.key.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.query.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __A : Optional[Any] = check_and_map_params( self_attn.query.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.value.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.value.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __A : BertSelfOutput = layer.attention.output __A : Tuple = check_and_map_params( self_output.dense.bias , F"""encoder.transformer_cells.{i}.proj.bias""" ) __A : int = check_and_map_params( self_output.dense.weight , F"""encoder.transformer_cells.{i}.proj.weight""" ) __A : List[Any] = check_and_map_params( self_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __A : str = check_and_map_params( self_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __A : BertIntermediate = layer.intermediate __A : int = check_and_map_params( intermediate.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __A : List[Any] = check_and_map_params( intermediate.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __A : BertOutput = layer.output __A : List[Any] = check_and_map_params( bert_output.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __A : Dict = check_and_map_params( bert_output.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __A : Optional[int] = check_and_map_params( bert_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __A : Dict = check_and_map_params( bert_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __A : Any = RobertaTokenizer.from_pretrained('roberta-base' ) __A : List[str] = tokenizer.encode_plus(a )['input_ids'] # Get gluon output __A : List[str] = mx.nd.array([input_ids] ) __A : Union[str, Any] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) __A : Optional[Any] = BertModel.from_pretrained(a ) hf_bort_model.eval() __A : Tuple = tokenizer.encode_plus(a , return_tensors='pt' ) __A : Any = hf_bort_model(**a )[0] __A : Union[str, Any] = output_gluon[0].asnumpy() __A : Tuple = output_hf[0].detach().numpy() __A : int = np.max(np.abs(hf_layer - gluon_layer ) ).item() __A : int = np.allclose(a , a , atol=1e-3 ) if success: print('✔️ Both model do output the same tensors' ) else: print('❌ Both model do **NOT** output the same tensors' ) print('Absolute difference is:' , a ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase : Dict = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
280
0
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS lowercase_ = logging.get_logger(__name__) lowercase_ = { '''linear''': get_linear_schedule_with_warmup, '''cosine''': get_cosine_schedule_with_warmup, '''cosine_w_restarts''': get_cosine_with_hard_restarts_schedule_with_warmup, '''polynomial''': get_polynomial_decay_schedule_with_warmup, '''constant''': get_constant_schedule, '''constant_w_warmup''': get_constant_schedule_with_warmup, } class __lowerCAmelCase ( snake_case__ ): def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , *lowerCAmelCase , **lowerCAmelCase ) -> int: '''simple docstring''' super().__init__(*_A , **_A ) if config is None: assert isinstance(self.model , _A ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F''' {self.model.__class__}''' ) _lowercase =self.model.config else: _lowercase =config _lowercase =data_args _lowercase =self.config.tgt_vocab_size if isinstance(self.config , _A ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' ' padding..' ) if self.args.label_smoothing == 0: _lowercase =torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _lowercase =label_smoothed_nll_loss def A__ ( self , lowerCAmelCase ) -> Any: '''simple docstring''' if self.optimizer is None: _lowercase =['bias', 'LayerNorm.weight'] _lowercase =[ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _lowercase =Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _lowercase =Adafactor _lowercase ={'scale_parameter': False, 'relative_step': False} else: _lowercase =AdamW _lowercase ={ 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _lowercase =self.args.learning_rate if self.sharded_ddp: _lowercase =OSS( params=_A , optim=_A , **_A , ) else: _lowercase =optimizer_cls(_A , **_A ) if self.lr_scheduler is None: _lowercase =self._get_lr_scheduler(_A ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def A__ ( self , lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowercase =arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _lowercase =schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _lowercase =schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: _lowercase =schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=_A ) return scheduler def A__ ( self ) -> Tuple: '''simple docstring''' if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Dict: '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _lowercase =model(**_A , use_cache=_A )[0] _lowercase =self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models _lowercase =model(**_A , labels=_A , use_cache=_A )[:2] else: # compute label smoothed loss _lowercase =model(**_A , use_cache=_A )[0] _lowercase =torch.nn.functional.log_softmax(_A , dim=-1 ) _lowercase =self.loss_fn(_A , _A , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def A__ ( self , lowerCAmelCase , lowerCAmelCase ) -> List[Any]: '''simple docstring''' _lowercase =inputs.pop('labels' ) _lowercase =self._compute_loss(_A , _A , _A ) return loss def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , ) -> Dict: '''simple docstring''' _lowercase =self._prepare_inputs(_A ) _lowercase ={ 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _lowercase =self.model.generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , **_A , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _lowercase =self._pad_tensors_to_max_len(_A , gen_kwargs['max_length'] ) _lowercase =inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _lowercase =self._compute_loss(_A , _A , _A ) _lowercase =loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _lowercase =generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _lowercase =self._pad_tensors_to_max_len(_A , gen_kwargs['max_length'] ) return (loss, logits, labels) def A__ ( self , lowerCAmelCase , lowerCAmelCase ) -> List[str]: '''simple docstring''' _lowercase =self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F''' padded to `max_length`={max_length}''' ) _lowercase =pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) _lowercase =tensor return padded_tensor
205
import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: __A : List[str] = x __A : str = y for step in range(a ): # noqa: B007 __A : Union[str, Any] = a * a - b * b + x __A : Optional[int] = 2 * a * b + y __A : List[str] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( a = 8_00 , a = 6_00 , a = -0.6 , a = 0 , a = 3.2 , a = 50 , a = True , ) -> Image.Image: __A : str = Image.new('RGB' , (image_width, image_height) ) __A : Dict = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates __A : Dict = figure_width / image_width * image_height __A : Union[str, Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width __A : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height __A : Union[str, Any] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __A : Optional[Any] = get_color_coded_rgb(a ) else: __A : Dict = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCAmelCase : str = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
280
0
'''simple docstring''' import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('.') def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( """`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got """ F'''{test_file} instead.''' ) UpperCAmelCase__ = components[-1] if not test_fn.endswith("""py""" ): raise ValueError(F'''`test_file` should be a python file. Got {test_fn} instead.''' ) if not test_fn.startswith("""test_modeling_""" ): raise ValueError( F'''`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.''' ) UpperCAmelCase__ = components[:-1] + [test_fn.replace(""".py""" , """""" )] UpperCAmelCase__ = '.'.join(SCREAMING_SNAKE_CASE__ ) return test_module_path def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' UpperCAmelCase__ = get_module_path(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = importlib.import_module(SCREAMING_SNAKE_CASE__ ) return test_module def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): '''simple docstring''' UpperCAmelCase__ = [] UpperCAmelCase__ = get_test_module(SCREAMING_SNAKE_CASE__ ) for attr in dir(SCREAMING_SNAKE_CASE__ ): if attr.endswith("""ModelTester""" ): tester_classes.append(getattr(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) # sort with class names return sorted(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : x.__name__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' UpperCAmelCase__ = [] UpperCAmelCase__ = get_test_module(SCREAMING_SNAKE_CASE__ ) for attr in dir(SCREAMING_SNAKE_CASE__ ): UpperCAmelCase__ = getattr(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). UpperCAmelCase__ = getattr(SCREAMING_SNAKE_CASE__ , """all_model_classes""" , [] ) if len(SCREAMING_SNAKE_CASE__ ) > 0: test_classes.append(SCREAMING_SNAKE_CASE__ ) # sort with class names return sorted(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : x.__name__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' UpperCAmelCase__ = get_test_classes(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : x.__name__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): '''simple docstring''' UpperCAmelCase__ = test_class() if hasattr(SCREAMING_SNAKE_CASE__ , """setUp""" ): test.setUp() UpperCAmelCase__ = None if hasattr(SCREAMING_SNAKE_CASE__ , """model_tester""" ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: UpperCAmelCase__ = test.model_tester.__class__ return model_tester def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Dict ): '''simple docstring''' UpperCAmelCase__ = get_test_classes(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(SCREAMING_SNAKE_CASE__ ) # sort with class names return sorted(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : x.__name__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' UpperCAmelCase__ = get_test_classes_for_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = [] for test_class in test_classes: UpperCAmelCase__ = get_model_tester_from_test_class(SCREAMING_SNAKE_CASE__ ) if tester_class is not None: tester_classes.append(SCREAMING_SNAKE_CASE__ ) # sort with class names return sorted(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : x.__name__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[str] ): '''simple docstring''' UpperCAmelCase__ = get_test_classes(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = {test_class: get_model_tester_from_test_class(SCREAMING_SNAKE_CASE__ ) for test_class in test_classes} return test_tester_mapping def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): '''simple docstring''' UpperCAmelCase__ = get_model_classes(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = { model_class: get_test_classes_for_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for model_class in model_classes } return model_test_mapping def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ = get_model_classes(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = { model_class: get_tester_classes_for_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for model_class in model_classes } return model_to_tester_mapping def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return o elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return o.__name__ elif isinstance(SCREAMING_SNAKE_CASE__ , (list, tuple) ): return [to_json(SCREAMING_SNAKE_CASE__ ) for x in o] elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return {to_json(SCREAMING_SNAKE_CASE__ ): to_json(SCREAMING_SNAKE_CASE__ ) for k, v in o.items()} else: return o
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE ( a , a , a , ) -> float: if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( a , nominal_annual_percentage_rate / 3_65 , number_of_years * 3_65 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self : List[str] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[Any]=7 ,lowerCamelCase__ : Dict=3 ,lowerCamelCase__ : Any=18 ,lowerCamelCase__ : List[str]=30 ,lowerCamelCase__ : int=400 ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : Any=None ,lowerCamelCase__ : str=True ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any=True ,): UpperCAmelCase__ = size if size is not None else {'shortest_edge': 20} UpperCAmelCase__ = crop_size if crop_size is not None else {'height': 18, 'width': 18} UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = min_resolution UpperCAmelCase__ = max_resolution UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = do_center_crop UpperCAmelCase__ = crop_size UpperCAmelCase__ = do_flip_channel_order def __lowerCAmelCase ( self : List[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" snake_case__ = MobileViTImageProcessor if is_vision_available() else None def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = MobileViTImageProcessingTester(self ) @property def __lowerCAmelCase ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : Any ): UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A ,'do_resize' ) ) self.assertTrue(hasattr(_A ,'size' ) ) self.assertTrue(hasattr(_A ,'do_center_crop' ) ) self.assertTrue(hasattr(_A ,'center_crop' ) ) self.assertTrue(hasattr(_A ,'do_flip_channel_order' ) ) def __lowerCAmelCase ( self : Dict ): UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'shortest_edge': 20} ) self.assertEqual(image_processor.crop_size ,{'height': 18, 'width': 18} ) UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ,crop_size=84 ) self.assertEqual(image_processor.size ,{'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size ,{'height': 84, 'width': 84} ) def __lowerCAmelCase ( self : Any ): pass def __lowerCAmelCase ( self : Union[str, Any] ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A ,Image.Image ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched UpperCAmelCase__ = image_processing(_A ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) def __lowerCAmelCase ( self : Dict ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_A ,numpify=_A ) for image in image_inputs: self.assertIsInstance(_A ,np.ndarray ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched UpperCAmelCase__ = image_processing(_A ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) def __lowerCAmelCase ( self : Tuple ): # Initialize image_processing UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_A ,torchify=_A ) for image in image_inputs: self.assertIsInstance(_A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched UpperCAmelCase__ = image_processing(_A ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,)
98
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
def __snake_case ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] ) -> int: A_ : int = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __snake_case ( _lowerCAmelCase : str , _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any] ) -> List[Any]: A_ : str = 0 while b > 0: if b & 1: A_ : str = ((res % c) + (a % c)) % c a += a b >>= 1 return res
300
def _SCREAMING_SNAKE_CASE ( a ) -> bool: return str(a ) == str(a )[::-1] def _SCREAMING_SNAKE_CASE ( a ) -> int: return int(a ) + int(str(a )[::-1] ) def _SCREAMING_SNAKE_CASE ( a = 1_00_00 ) -> int: __A : int = [] for num in range(1 , a ): __A : List[str] = 0 __A : List[Any] = num while iterations < 50: __A : str = sum_reverse(a ) iterations += 1 if is_palindrome(a ): break else: lychrel_nums.append(a ) return len(a ) if __name__ == "__main__": print(F"""{solution() = }""")
280
0
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowercase_ = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize lowercase_ = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' lowercase_ = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' lowercase_ = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): '''simple docstring''' def snake_case_( self ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"""] , reference_urls=[ """https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score""", """https://en.wikipedia.org/wiki/METEOR""", ] , ) def snake_case_( self , A ) -> Optional[int]: import nltk nltk.download("""wordnet""" ) if NLTK_VERSION >= version.Version("""3.6.5""" ): nltk.download("""punkt""" ) if NLTK_VERSION >= version.Version("""3.6.6""" ): nltk.download("""omw-1.4""" ) def snake_case_( self , A , A , A=0.9 , A=3 , A=0.5 ) -> List[str]: if NLTK_VERSION >= version.Version("""3.6.5""" ): _SCREAMING_SNAKE_CASE = [ meteor_score.single_meteor_score( word_tokenize(_A ) , word_tokenize(_A ) , alpha=_A , beta=_A , gamma=_A ) for ref, pred in zip(_A , _A ) ] else: _SCREAMING_SNAKE_CASE = [ meteor_score.single_meteor_score(_A , _A , alpha=_A , beta=_A , gamma=_A ) for ref, pred in zip(_A , _A ) ] return {"meteor": np.mean(_A )}
58
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class _A: """simple docstring""" def __init__( self , _A = None ): if components is None: __A : int = [] __A : Tuple = list(_A ) def __len__( self ): return len(self.__components ) def __str__( self ): return "(" + ",".join(map(_A , self.__components ) ) + ")" def __add__( self , _A ): __A : Optional[int] = len(self ) if size == len(_A ): __A : Any = [self.__components[i] + other.component(_A ) for i in range(_A )] return Vector(_A ) else: raise Exception('must have the same size' ) def __sub__( self , _A ): __A : Tuple = len(self ) if size == len(_A ): __A : Union[str, Any] = [self.__components[i] - other.component(_A ) for i in range(_A )] return Vector(_A ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , (float, int) ): __A : str = [c * other for c in self.__components] return Vector(_A ) elif isinstance(_A , _A ) and len(self ) == len(_A ): __A : Union[str, Any] = len(self ) __A : Dict = [self.__components[i] * other.component(_A ) for i in range(_A )] return sum(_A ) else: # error case raise Exception('invalid operand!' ) def UpperCAmelCase_ ( self ): return Vector(self.__components ) def UpperCAmelCase_ ( self , _A ): if isinstance(_A , _A ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def UpperCAmelCase_ ( self , _A , _A ): assert -len(self.__components ) <= pos < len(self.__components ) __A : Optional[int] = value def UpperCAmelCase_ ( self ): if len(self.__components ) == 0: raise Exception('Vector is empty' ) __A : Optional[Any] = [c**2 for c in self.__components] return math.sqrt(sum(_A ) ) def UpperCAmelCase_ ( self , _A , _A = False ): __A : Optional[Any] = self * other __A : Optional[Any] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def _SCREAMING_SNAKE_CASE ( a ) -> Vector: assert isinstance(a , a ) return Vector([0] * dimension ) def _SCREAMING_SNAKE_CASE ( a , a ) -> Vector: assert isinstance(a , a ) and (isinstance(a , a )) __A : Optional[Any] = [0] * dimension __A : Tuple = 1 return Vector(a ) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: assert ( isinstance(a , a ) and isinstance(a , a ) and (isinstance(a , (int, float) )) ) return x * scalar + y def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Vector: random.seed(a ) __A : str = [random.randint(a , a ) for _ in range(a )] return Vector(a ) class _A: """simple docstring""" def __init__( self , _A , _A , _A ): __A : Optional[Any] = matrix __A : Dict = w __A : Optional[int] = h def __str__( self ): __A : Tuple = '' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Optional[Any] = [] for i in range(self.__height ): __A : Optional[Any] = [ self.__matrix[i][j] + other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self , _A ): if self.__width == other.width() and self.__height == other.height(): __A : Tuple = [] for i in range(self.__height ): __A : str = [ self.__matrix[i][j] - other.component(_A , _A ) for j in range(self.__width ) ] matrix.append(_A ) return Matrix(_A , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self , _A ): ... @overload def __mul__( self , _A ): ... def __mul__( self , _A ): if isinstance(_A , _A ): # matrix-vector if len(_A ) == self.__width: __A : List[Any] = zero_vector(self.__height ) for i in range(self.__height ): __A : List[str] = [ self.__matrix[i][j] * other.component(_A ) for j in range(self.__width ) ] ans.change_component(_A , sum(_A ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(_A , (int, float) ): # matrix-scalar __A : List[str] = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(_A , self.__width , self.__height ) return None def UpperCAmelCase_ ( self ): return self.__height def UpperCAmelCase_ ( self ): return self.__width def UpperCAmelCase_ ( self , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A , _A ): if 0 <= x < self.__height and 0 <= y < self.__width: __A : int = value else: raise Exception('change_component: indices out of bounds' ) def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) __A : List[str] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(_A ) ): __A : Optional[int] = minor[i][:y] + minor[i][y + 1 :] return Matrix(_A , self.__width - 1 , self.__height - 1 ).determinant() def UpperCAmelCase_ ( self , _A , _A ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(_A , _A ) else: raise Exception('Indices out of bounds' ) def UpperCAmelCase_ ( self ): if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __A : List[str] = [ self.__matrix[0][y] * self.cofactor(0 , _A ) for y in range(self.__width ) ] return sum(_A ) def _SCREAMING_SNAKE_CASE ( a ) -> Matrix: __A : list[list[float]] = [[0] * n for _ in range(a )] return Matrix(a , a , a ) def _SCREAMING_SNAKE_CASE ( a , a , a , a ) -> Matrix: random.seed(a ) __A : list[list[float]] = [ [random.randint(a , a ) for _ in range(a )] for _ in range(a ) ] return Matrix(a , a , a )
280
0
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : List[str] = [] for line in lines: SCREAMING_SNAKE_CASE : List[str] = re.sub(r"#.*" , "" , _a) # remove comments if line: filtered_lines.append(_a) SCREAMING_SNAKE_CASE : List[str] = '\n'.join(_a) # Make a hash from all this code SCREAMING_SNAKE_CASE : str = full_str.encode("utf-8") return shaaaa(_a).hexdigest() # get importable module names and hash for caching a_ = { '''csv''': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), '''json''': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), '''pandas''': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), '''parquet''': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), '''arrow''': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), '''text''': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), '''imagefolder''': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), '''audiofolder''': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions a_ = { '''.csv''': ('''csv''', {}), '''.tsv''': ('''csv''', {'''sep''': '''\t'''}), '''.json''': ('''json''', {}), '''.jsonl''': ('''json''', {}), '''.parquet''': ('''parquet''', {}), '''.arrow''': ('''arrow''', {}), '''.txt''': ('''text''', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) a_ = {'''imagefolder''', '''audiofolder'''} # Used to filter data files based on extensions given a module name a_ = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
76
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase : List[str] = '''▁''' UpperCAmelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Optional[int] = BertGenerationTokenizer UpperCamelCase : str = False UpperCamelCase : Tuple = True def UpperCAmelCase_ ( self ): super().setUp() __A : Tuple = BertGenerationTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : str = '<s>' __A : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def UpperCAmelCase_ ( self ): __A : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(_A ) , 1002 ) def UpperCAmelCase_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCAmelCase_ ( self ): __A : str = BertGenerationTokenizer(_A , keep_accents=_A ) __A : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_A , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [285, 46, 10, 170, 382] , ) __A : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __A : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __A : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase_ ( self ): return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def UpperCAmelCase_ ( self ): __A : List[Any] = 'Hello World!' __A : Optional[Any] = [18536, 2260, 101] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def UpperCAmelCase_ ( self ): __A : Dict = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) __A : int = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @require_torch @slow def UpperCAmelCase_ ( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence __A : Tuple = list(self.big_tokenizer.get_vocab().keys() )[:10] __A : List[Any] = ' '.join(_A ) __A : Union[str, Any] = self.big_tokenizer.encode_plus(_A , return_tensors='pt' , return_token_type_ids=_A ) __A : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=_A ) __A : int = BertGenerationConfig() __A : List[str] = BertGenerationEncoder(_A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def UpperCAmelCase_ ( self ): # fmt: off __A : str = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
280
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/config.json''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/config.json''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json''' ), } class A ( snake_case__ ): __UpperCAmelCase : str = '''xlm-roberta''' def __init__(self : Optional[Any] , __UpperCAmelCase : List[str]=3_0_5_2_2 , __UpperCAmelCase : Tuple=7_6_8 , __UpperCAmelCase : Optional[Any]=1_2 , __UpperCAmelCase : List[Any]=1_2 , __UpperCAmelCase : int=3_0_7_2 , __UpperCAmelCase : Optional[int]="gelu" , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : Optional[int]=0.1 , __UpperCAmelCase : str=5_1_2 , __UpperCAmelCase : Any=2 , __UpperCAmelCase : Optional[int]=0.02 , __UpperCAmelCase : Optional[Any]=1E-12 , __UpperCAmelCase : int=1 , __UpperCAmelCase : Tuple=0 , __UpperCAmelCase : List[str]=2 , __UpperCAmelCase : Optional[Any]="absolute" , __UpperCAmelCase : str=True , __UpperCAmelCase : List[Any]=None , **__UpperCAmelCase : str , ) -> List[Any]: """simple docstring""" super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_act UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = position_embedding_type UpperCAmelCase__ = use_cache UpperCAmelCase__ = classifier_dropout class A ( snake_case__ ): @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase__ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: UpperCAmelCase__ = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
65
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _A: """simple docstring""" @staticmethod def UpperCAmelCase_ ( *_A , **_A ): pass def _SCREAMING_SNAKE_CASE ( a ) -> str: __A : str = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def _SCREAMING_SNAKE_CASE ( a ) -> Dict: __A : Dict = np.array(a ) __A : List[Any] = npimg.shape return {"hash": hashimage(a ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _A( unittest.TestCase ): """simple docstring""" UpperCamelCase : str = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase : int = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def UpperCAmelCase_ ( self , _A , _A , _A ): __A : Dict = MaskGenerationPipeline(model=_A , image_processor=_A ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase_ ( self , _A , _A ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def UpperCAmelCase_ ( self ): pass @slow @require_torch def UpperCAmelCase_ ( self ): __A : Union[str, Any] = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) __A : List[str] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=256 ) # Shortening by hashing __A : List[Any] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] , ) # fmt: on @require_torch @slow def UpperCAmelCase_ ( self ): __A : Optional[Any] = 'facebook/sam-vit-huge' __A : List[str] = pipeline('mask-generation' , model=_A ) __A : Tuple = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __A : List[str] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(_A ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_A , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] , )
280
0
"""simple docstring""" import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: int = 'hf-internal-testing/tiny-random-t5' a__: List[str] = AutoTokenizer.from_pretrained(_A) a__: Dict = AutoModelForSeqaSeqLM.from_pretrained(_A) a__: str = tokenizer('This is me' , return_tensors='pt') a__: str = model.to_bettertransformer() self.assertTrue(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules())) a__: Any = model.generate(**_A) a__: Any = model.reverse_bettertransformer() self.assertFalse(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules())) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_A) a__: Dict = AutoModelForSeqaSeqLM.from_pretrained(_A) self.assertFalse( any('BetterTransformer' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules())) a__: Optional[Any] = model_reloaded.generate(**_A) self.assertTrue(torch.allclose(_A , _A)) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: List[Any] = 'hf-internal-testing/tiny-random-t5' a__: List[Any] = AutoModelForSeqaSeqLM.from_pretrained(_A) a__: Tuple = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(_A): model.save_pretrained(_A) a__: List[Any] = model.reverse_bettertransformer() model.save_pretrained(_A)
290
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): __A : List[Any] = tempfile.mkdtemp() # fmt: off __A : List[str] = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on __A : Union[str, Any] = dict(zip(_A , range(len(_A ) ) ) ) __A : Optional[int] = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] __A : int = {'unk_token': '<unk>'} __A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_A ) ) __A : List[Any] = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], 'image_std': [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } __A : Optional[int] = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **_A ) def UpperCAmelCase_ ( self , **_A ): return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def UpperCAmelCase_ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self ): __A : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __A : Optional[int] = [Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase_ ( self ): __A : List[Any] = self.get_tokenizer() __A : str = self.get_rust_tokenizer() __A : List[str] = self.get_image_processor() __A : Optional[int] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_slow.save_pretrained(self.tmpdirname ) __A : int = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) processor_fast.save_pretrained(self.tmpdirname ) __A : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _A ) self.assertIsInstance(processor_fast.tokenizer , _A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _A ) self.assertIsInstance(processor_fast.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : List[str] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) __A : Optional[int] = self.get_image_processor(do_normalize=_A ) __A : Any = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Optional[Any] = self.get_tokenizer() __A : Union[str, Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Union[str, Any] = self.prepare_image_inputs() __A : int = image_processor(_A , return_tensors='np' ) __A : str = processor(images=_A , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCAmelCase_ ( self ): __A : str = self.get_image_processor() __A : str = self.get_tokenizer() __A : Tuple = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : str = 'lower newer' __A : str = processor(text=_A , return_tensors='np' ) __A : List[str] = tokenizer(_A , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCAmelCase_ ( self ): __A : int = self.get_image_processor() __A : Optional[int] = self.get_tokenizer() __A : List[str] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Any = 'lower newer' __A : Optional[Any] = self.prepare_image_inputs() __A : List[Any] = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Any = 'google/owlvit-base-patch32' __A : int = OwlViTProcessor.from_pretrained(_A ) __A : Dict = ['cat', 'nasa badge'] __A : Optional[Any] = processor(text=_A ) __A : Optional[int] = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Tuple = 'google/owlvit-base-patch32' __A : Any = OwlViTProcessor.from_pretrained(_A ) __A : Dict = [['cat', 'nasa badge'], ['person']] __A : Dict = processor(text=_A ) __A : Optional[int] = 16 __A : Any = len(_A ) __A : Union[str, Any] = max([len(_A ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : List[Any] = 'google/owlvit-base-patch32' __A : str = OwlViTProcessor.from_pretrained(_A ) __A : Union[str, Any] = ['cat', 'nasa badge'] __A : Tuple = processor(text=_A ) __A : str = 16 __A : int = inputs['input_ids'] __A : List[Any] = [ [49406, 2368, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49406, 6841, 11301, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : List[str] = self.get_tokenizer() __A : Optional[Any] = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = self.prepare_image_inputs() __A : Optional[int] = processor(images=_A , query_images=_A ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def UpperCAmelCase_ ( self ): __A : Optional[Any] = self.get_image_processor() __A : Union[str, Any] = self.get_tokenizer() __A : str = OwlViTProcessor(tokenizer=_A , image_processor=_A ) __A : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A : Any = processor.batch_decode(_A ) __A : Tuple = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A )
280
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} SCREAMING_SNAKE_CASE__ : str = { '''vocab_file''': { '''squeezebert/squeezebert-uncased''': ( '''https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt''' ), '''squeezebert/squeezebert-mnli''': '''https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt''', '''squeezebert/squeezebert-mnli-headless''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''squeezebert/squeezebert-uncased''': ( '''https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json''' ), '''squeezebert/squeezebert-mnli''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json''' ), '''squeezebert/squeezebert-mnli-headless''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json''' ), }, } SCREAMING_SNAKE_CASE__ : Tuple = { '''squeezebert/squeezebert-uncased''': 512, '''squeezebert/squeezebert-mnli''': 512, '''squeezebert/squeezebert-mnli-headless''': 512, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { '''squeezebert/squeezebert-uncased''': {'''do_lower_case''': True}, '''squeezebert/squeezebert-mnli''': {'''do_lower_case''': True}, '''squeezebert/squeezebert-mnli-headless''': {'''do_lower_case''': True}, } class UpperCamelCase__ (snake_case__ ): '''simple docstring''' lowerCamelCase_ : str = VOCAB_FILES_NAMES lowerCamelCase_ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ : List[str] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ : Optional[int] = SqueezeBertTokenizer def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=True , UpperCamelCase__="[UNK]" , UpperCamelCase__="[SEP]" , UpperCamelCase__="[PAD]" , UpperCamelCase__="[CLS]" , UpperCamelCase__="[MASK]" , UpperCamelCase__=True , UpperCamelCase__=None , **UpperCamelCase__ , ) -> int: super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , tokenize_chinese_chars=_A , strip_accents=_A , **_A , ) lowerCamelCase : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , _A ) != do_lower_case or normalizer_state.get("strip_accents" , _A ) != strip_accents or normalizer_state.get("handle_chinese_chars" , _A ) != tokenize_chinese_chars ): lowerCamelCase : Any = getattr(_A , normalizer_state.pop("type" ) ) lowerCamelCase : Union[str, Any] = do_lower_case lowerCamelCase : List[str] = strip_accents lowerCamelCase : Optional[int] = tokenize_chinese_chars lowerCamelCase : Union[str, Any] = normalizer_class(**_A ) lowerCamelCase : Dict = do_lower_case def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__=None ) -> Optional[Any]: lowerCamelCase : int = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[Any]: lowerCamelCase : Optional[int] = [self.sep_token_id] lowerCamelCase : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Any: lowerCamelCase : List[str] = self._tokenizer.model.save(_A , name=_A ) return tuple(_A )
48
import math def _SCREAMING_SNAKE_CASE ( a ) -> list[int]: __A : List[str] = [] __A : Any = 2 __A : Union[str, Any] = int(math.sqrt(a ) ) # Size of every segment __A : Any = [True] * (end + 1) __A : List[Any] = [] while start <= end: if temp[start] is True: in_prime.append(a ) for i in range(start * start , end + 1 , a ): __A : Optional[int] = False start += 1 prime += in_prime __A : Any = end + 1 __A : Any = min(2 * end , a ) while low <= n: __A : List[Any] = [True] * (high - low + 1) for each in in_prime: __A : List[str] = math.floor(low / each ) * each if t < low: t += each for j in range(a , high + 1 , a ): __A : Optional[int] = False for j in range(len(a ) ): if temp[j] is True: prime.append(j + low ) __A : Optional[int] = high + 1 __A : Tuple = min(high + end , a ) return prime print(sieve(10**6))
280
0
import enum import shutil import sys lowercase = shutil.get_terminal_size() lowercase = {'''UP''': '''A''', '''DOWN''': '''B''', '''RIGHT''': '''C''', '''LEFT''': '''D'''} class UpperCamelCase_ ( enum.Enum ): '''simple docstring''' lowerCAmelCase = 0 lowerCAmelCase = 1 def __UpperCAmelCase ( a_ , a_=""): sys.stdout.write(str(a_) + end) sys.stdout.flush() def __UpperCAmelCase ( a_ , a_ , a_=""): forceWrite(f'''\u001b[{color}m{content}\u001b[0m''' , a_) def __UpperCAmelCase ( ): forceWrite('\r') def __UpperCAmelCase ( a_ , a_): forceWrite(f'''\033[{num_lines}{CURSOR_TO_CHAR[direction.upper()]}''') def __UpperCAmelCase ( ): forceWrite(' ' * TERMINAL_WIDTH) reset_cursor() def __UpperCAmelCase ( ): reset_cursor() forceWrite('-' * TERMINAL_WIDTH)
178
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : str = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): UpperCamelCase_ = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right UpperCamelCase_ = 12_80_22 UpperCamelCase_ = 12_80_28 @require_sentencepiece class a_ (snake_case__ , unittest.TestCase ): __lowerCAmelCase : Optional[int] = MaMaaaTokenizer __lowerCAmelCase : Optional[Any] = False __lowerCAmelCase : str = False __lowerCAmelCase : Any = True def __UpperCamelCase ( self ): super().setUp() _lowerCAmelCase : int = ['</s>', '<unk>', '▁This', '▁is', '▁a', '▁t', 'est', '\u0120', '<pad>'] _lowerCAmelCase : Tuple = dict(zip(_A , range(len(_A ) ) ) ) _lowerCAmelCase : List[Any] = Path(self.tmpdirname ) save_json(_A , save_dir / VOCAB_FILES_NAMES["""vocab_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(_A , save_dir / VOCAB_FILES_NAMES["""spm_file"""] ) _lowerCAmelCase : Optional[int] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCamelCase ( self , **snake_case_ ): return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **_A ) def __UpperCamelCase ( self , snake_case_ ): return ( "This is a test", "This is a test", ) def __UpperCamelCase ( self ): _lowerCAmelCase : Tuple = '</s>' _lowerCAmelCase : Dict = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = self.get_tokenizer() _lowerCAmelCase : Dict = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<s>""" ) self.assertEqual(len(_A ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("""Skip this test while all models are still to be uploaded.""" ) def __UpperCamelCase ( self ): pass def __UpperCamelCase ( self ): _lowerCAmelCase : List[str] = self.get_tokenizer() _lowerCAmelCase : List[str] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(_A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [2, 3, 4, 5, 6] , ) _lowerCAmelCase : Optional[int] = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(_A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) _lowerCAmelCase : str = tokenizer.convert_tokens_to_string(_A ) self.assertEqual(_A , """This is a test""" ) @slow def __UpperCamelCase ( self ): # fmt: off _lowerCAmelCase : List[Any] = {'input_ids': [[1_2_8_0_2_2, 1_1_0_1_0_8, 3_9_7, 1_1, 3_8_2_7_2, 2_2_4_7, 1_2_4_8_1_1, 2_8_5, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 3_9_5_3_4, 4_4_2_8, 3_9_7, 1_0_1_9, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 4_1_3_3_7, 1_6_7_8_6, 2_4_1, 7, 2_0_2_1_4, 1_7, 1_2_5_6_9_0, 1_0_3_9_8, 7, 4_4_3_7_8, 5_8_0_6_9, 6_8_3_4_2, 7_7_9_8, 7_3_4_3, 1_1, 2_9_9, 3_3_3_1_0, 4, 1_5_8, 3_7_3_5_0, 9_4_0_7_7, 4_5_6_9, 2_9_9, 3_3_3_1_0, 9_0, 4, 5_2_8_4_0, 2_9_0, 4, 3_1_2_7_0, 1_1_2, 2_9_9, 6_8_2, 4, 5_2_8_4_0, 3_9_9_5_3, 1_4_0_7_9, 1_9_3, 5_2_5_1_9, 9_0_8_9_4, 1_7_8_9_4, 1_2_0_6_9_7, 1_1, 4_0_4_4_5, 5_5_1, 1_7, 1_0_1_9, 5_2_5_1_9, 9_0_8_9_4, 1_7_7_5_6, 9_6_3, 1_1, 4_0_4_4_5, 4_8_0, 1_7, 9_7_9_2, 1_1_2_0, 5_1_7_3, 1_3_9_3, 6_2_4_0, 1_6_7_8_6, 2_4_1, 1_2_0_9_9_6, 2_8, 1_2_4_5, 1_3_9_3, 1_1_8_2_4_0, 1_1_1_2_3, 1_0_1_9, 9_3_6_1_2, 2_6_9_1, 1_0_6_1_8, 9_8_0_5_8, 1_2_0_4_0_9, 1_9_2_8, 2_7_9, 4, 4_0_6_8_3, 3_6_7, 1_7_8, 2_0_7, 1_0_1_9, 1_0_3, 1_0_3_1_2_1, 5_0_6, 6_5_2_9_6, 5, 2], [1_2_8_0_2_2, 2_1_2_1_7, 3_6_7, 1_1_7, 1_2_5_4_5_0, 1_2_8, 7_1_9, 7, 7_3_0_8, 4_0, 9_3_6_1_2, 1_2_6_6_9, 1_1_1_6, 1_6_7_0_4, 7_1, 1_7_7_8_5, 3_6_9_9, 1_5_5_9_2, 3_5, 1_4_4, 9_5_8_4, 2_4_1, 1_1_9_4_3, 7_1_3, 9_5_0, 7_9_9, 2_2_4_7, 8_8_4_2_7, 1_5_0, 1_4_9, 1_1_8_8_1_3, 1_2_0_7_0_6, 1_0_1_9, 1_0_6_9_0_6, 8_1_5_1_8, 2_8, 1_2_2_4, 2_2_7_9_9, 3_9_7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1_2_8_0_2_2, 1_6_5_8, 1_2_3_3_1_1, 5_1_5_5, 5_5_7_8, 4_7_2_2, 2_7_9, 1_4_9_4_7, 2_3_6_6, 1_1_2_0, 1_1_9_7, 1_4, 1_3_4_8, 9_2_3_2, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name="""facebook/m2m100_418M""" , revision="""c168bae485c864188cf9aa0e4108b0b6934dc91e""" , ) @require_torch @require_sentencepiece @require_tokenizers class a_ (unittest.TestCase ): __lowerCAmelCase : List[Any] = '''facebook/m2m100_418M''' __lowerCAmelCase : Any = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] __lowerCAmelCase : Optional[Any] = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off __lowerCAmelCase : List[str] = [EN_CODE, 5_9_3, 1_9_4_9, 1_1_5_7_8_1, 4, 7_1_5_8_6, 4_2_3_4, 6_0_6_3_3, 1_2_6_2_3_3, 4_3_2, 1_2_3_8_0_8, 1_5_5_9_2, 1_1_9_7, 1_1_7_1_3_2, 1_2_0_6_1_8, 5, 2] @classmethod def __UpperCamelCase ( cls ): _lowerCAmelCase : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en""" , tgt_lang="""fr""" ) _lowerCAmelCase : Optional[Any] = 1 return cls def __UpperCamelCase ( self ): self.assertEqual(self.tokenizer.get_lang_id("""ar""" ) , 1_2_8_0_0_6 ) self.assertEqual(self.tokenizer.get_lang_id("""en""" ) , 1_2_8_0_2_2 ) self.assertEqual(self.tokenizer.get_lang_id("""ro""" ) , 1_2_8_0_7_6 ) self.assertEqual(self.tokenizer.get_lang_id("""mr""" ) , 1_2_8_0_6_3 ) def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = self.tokenizer.get_vocab() self.assertEqual(len(_A ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["""<unk>"""] , 3 ) self.assertIn(self.tokenizer.get_lang_token("""en""" ) , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : Any = 'en' _lowerCAmelCase : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _A ) def __UpperCamelCase ( self ): self.assertIn(_A , self.tokenizer.all_special_ids ) # fmt: off _lowerCAmelCase : str = [FR_CODE, 5_3_6_4, 8_2, 8_6_4_2, 4, 2_9_4, 4_7, 8, 1_4_0_2_8, 1_3_6, 3_2_8_6, 9_7_0_6, 6, 9_0_7_9_7, 6, 1_4_4_0_1_2, 1_6_2, 8_8_1_2_8, 3_0_0_6_1, 5, 2] # fmt: on _lowerCAmelCase : str = self.tokenizer.decode(_A , skip_special_tokens=_A ) _lowerCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_A ) self.assertEqual(_A , _A ) self.assertNotIn(self.tokenizer.eos_token , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : int = tempfile.mkdtemp() _lowerCAmelCase : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(_A ) _lowerCAmelCase : Union[str, Any] = MaMaaaTokenizer.from_pretrained(_A ) self.assertDictEqual(new_tok.lang_token_to_id , _A ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = 'en' _lowerCAmelCase : Tuple = 'fr' _lowerCAmelCase : List[str] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_A , return_tensors="""pt""" ) _lowerCAmelCase : Union[str, Any] = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _lowerCAmelCase : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : Tuple = 'mr' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _lowerCAmelCase : List[Any] = 'zh' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : int = 'mr' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _lowerCAmelCase : Optional[Any] = 'zh' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : Any = self.tokenizer._build_translation_inputs("""A test""" , return_tensors="""pt""" , src_lang="""en""" , tgt_lang="""ar""" ) self.assertEqual( nested_simplify(_A ) , { # en_XX, A, test, EOS """input_ids""": [[1_2_8_0_2_2, 5_8, 4_1_8_3, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 1_2_8_0_0_6, } , )
309
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class __lowerCAmelCase ( snake_case__ ): _a = 42 class __lowerCAmelCase ( snake_case__ , snake_case__ ): @register_to_config def __init__( self , lowerCAmelCase = 3 , lowerCAmelCase = 3 , lowerCAmelCase = ("DownEncoderBlock2D",) , lowerCAmelCase = ("UpDecoderBlock2D",) , lowerCAmelCase = (64,) , lowerCAmelCase = 1 , lowerCAmelCase = "silu" , lowerCAmelCase = 3 , lowerCAmelCase = 32 , lowerCAmelCase = 256 , lowerCAmelCase = 32 , lowerCAmelCase = None , lowerCAmelCase = 0.18215 , lowerCAmelCase = "group" , ) -> int: '''simple docstring''' super().__init__() # pass init params to Encoder _lowercase =Encoder( in_channels=_A , out_channels=_A , down_block_types=_A , block_out_channels=_A , layers_per_block=_A , act_fn=_A , norm_num_groups=_A , double_z=_A , ) _lowercase =vq_embed_dim if vq_embed_dim is not None else latent_channels _lowercase =nn.Convad(_A , _A , 1 ) _lowercase =VectorQuantizer(_A , _A , beta=0.25 , remap=_A , sane_index_shape=_A ) _lowercase =nn.Convad(_A , _A , 1 ) # pass init params to Decoder _lowercase =Decoder( in_channels=_A , out_channels=_A , up_block_types=_A , block_out_channels=_A , layers_per_block=_A , act_fn=_A , norm_num_groups=_A , norm_type=_A , ) @apply_forward_hook def A__ ( self , lowerCAmelCase , lowerCAmelCase = True ) -> Any: '''simple docstring''' _lowercase =self.encoder(_A ) _lowercase =self.quant_conv(_A ) if not return_dict: return (h,) return VQEncoderOutput(latents=_A ) @apply_forward_hook def A__ ( self , lowerCAmelCase , lowerCAmelCase = False , lowerCAmelCase = True ) -> Union[str, Any]: '''simple docstring''' if not force_not_quantize: _lowercase =self.quantize(_A ) else: _lowercase =h _lowercase =self.post_quant_conv(_A ) _lowercase =self.decoder(_A , quant if self.config.norm_type == 'spatial' else None ) if not return_dict: return (dec,) return DecoderOutput(sample=_A ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = True ) -> List[str]: '''simple docstring''' _lowercase =sample _lowercase =self.encode(_A ).latents _lowercase =self.decode(_A ).sample if not return_dict: return (dec,) return DecoderOutput(sample=_A )
205
def _SCREAMING_SNAKE_CASE ( a , a = 0 ) -> list: __A : int = length or len(a ) __A : str = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __A , __A : Optional[int] = list_data[i + 1], list_data[i] __A : Union[str, Any] = True return list_data if not swapped else bubble_sort(a , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { '''facebook/levit-128S''': '''https://huggingface.co/facebook/levit-128S/resolve/main/config.json''', # See all LeViT models at https://huggingface.co/models?filter=levit } class lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' lowerCAmelCase_ : int = '''levit''' def __init__( self : int , _UpperCAmelCase : str=2_24 , _UpperCAmelCase : str=3 , _UpperCAmelCase : Tuple=3 , _UpperCAmelCase : Union[str, Any]=2 , _UpperCAmelCase : Dict=1 , _UpperCAmelCase : List[str]=16 , _UpperCAmelCase : Any=[1_28, 2_56, 3_84] , _UpperCAmelCase : int=[4, 8, 12] , _UpperCAmelCase : Optional[int]=[4, 4, 4] , _UpperCAmelCase : List[str]=[16, 16, 16] , _UpperCAmelCase : List[Any]=0 , _UpperCAmelCase : Optional[int]=[2, 2, 2] , _UpperCAmelCase : List[str]=[2, 2, 2] , _UpperCAmelCase : List[str]=0.02 , **_UpperCAmelCase : str , ): """simple docstring""" super().__init__(**_A ) UpperCAmelCase__ = image_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = kernel_size UpperCAmelCase__ = stride UpperCAmelCase__ = padding UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = depths UpperCAmelCase__ = key_dim UpperCAmelCase__ = drop_path_rate UpperCAmelCase__ = patch_size UpperCAmelCase__ = attention_ratio UpperCAmelCase__ = mlp_ratio UpperCAmelCase__ = initializer_range UpperCAmelCase__ = [ ['Subsample', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['Subsample', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] class lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' lowerCAmelCase_ : int = version.parse("""1.11""" ) @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return 1E-4
346
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( a ) -> int: if not nums: return 0 __A : Optional[int] = nums[0] __A : str = 0 for num in nums[1:]: __A , __A : Tuple = ( max_excluding + num, max(a , a ), ) return max(a , a ) if __name__ == "__main__": import doctest doctest.testmod()
280
0
"""simple docstring""" import argparse import json from tqdm import tqdm def a_ ( ): UpperCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=lowerCamelCase , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=lowerCamelCase , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=lowerCamelCase , help='where to store parsed gold_data_path file' , ) UpperCAmelCase__ = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: UpperCAmelCase__ = json.load(lowerCamelCase ) for dpr_record in tqdm(lowerCamelCase ): UpperCAmelCase__ = dpr_record['question'] UpperCAmelCase__ = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(lowerCamelCase ) + '\n' ) if __name__ == "__main__": main()
98
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
import argparse import torch from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert from transformers.utils import logging logging.set_verbosity_info() def __snake_case ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> Union[str, Any]: # Initialise PyTorch model A_ : Dict = BertConfig.from_json_file(_lowerCAmelCase ) print(f"Building PyTorch model from configuration: {config}" ) A_ : str = BertForPreTraining(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_bert(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _lowerCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--bert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _lowerCAmelCase : Tuple = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
300
def _SCREAMING_SNAKE_CASE ( a ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) __A : Optional[int] = len(bin(a )[3:] ) __A : Dict = bin(abs(a ) - (1 << binary_number_length) )[3:] __A : int = ( ( '1' + '0' * (binary_number_length - len(a )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
280
0
'''simple docstring''' from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar lowercase_ = TypeVar("""KEY""") lowercase_ = TypeVar("""VAL""") @dataclass(frozen=snake_case__ , slots=snake_case__ ) class a_ ( Generic[KEY, VAL] ): '''simple docstring''' UpperCamelCase = 42 UpperCamelCase = 42 class a_ ( _Item ): '''simple docstring''' def __init__( self ) -> List[str]: super().__init__(_A , _A ) def __bool__( self ) -> Any: return False lowercase_ = _DeletedItem() class a_ ( MutableMapping[KEY, VAL] ): '''simple docstring''' def __init__( self , A = 8 , A = 0.75 ) -> int: _SCREAMING_SNAKE_CASE = initial_block_size _SCREAMING_SNAKE_CASE = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 _SCREAMING_SNAKE_CASE = capacity_factor _SCREAMING_SNAKE_CASE = 0 def snake_case_( self , A ) -> List[Any]: return hash(_A ) % len(self._buckets ) def snake_case_( self , A ) -> Union[str, Any]: return (ind + 1) % len(self._buckets ) def snake_case_( self , A , A , A ) -> Optional[Any]: _SCREAMING_SNAKE_CASE = self._buckets[ind] if not stored: _SCREAMING_SNAKE_CASE = _Item(_A , _A ) self._len += 1 return True elif stored.key == key: _SCREAMING_SNAKE_CASE = _Item(_A , _A ) return True else: return False def snake_case_( self ) -> Dict: _SCREAMING_SNAKE_CASE = len(self._buckets ) * self._capacity_factor return len(self ) >= int(_A ) def snake_case_( self ) -> Optional[Any]: if len(self._buckets ) <= self._initial_block_size: return False _SCREAMING_SNAKE_CASE = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def snake_case_( self , A ) -> str: _SCREAMING_SNAKE_CASE = self._buckets _SCREAMING_SNAKE_CASE = [None] * new_size _SCREAMING_SNAKE_CASE = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def snake_case_( self ) -> List[str]: self._resize(len(self._buckets ) * 2 ) def snake_case_( self ) -> str: self._resize(len(self._buckets ) // 2 ) def snake_case_( self , A ) -> Dict: _SCREAMING_SNAKE_CASE = self._get_bucket_index(_A ) for _ in range(len(self._buckets ) ): yield ind _SCREAMING_SNAKE_CASE = self._get_next_ind(_A ) def snake_case_( self , A , A ) -> int: for ind in self._iterate_buckets(_A ): if self._try_set(_A , _A , _A ): break def __setitem__( self , A , A ) -> Tuple: if self._is_full(): self._size_up() self._add_item(_A , _A ) def __delitem__( self , A ) -> List[str]: for ind in self._iterate_buckets(_A ): _SCREAMING_SNAKE_CASE = self._buckets[ind] if item is None: raise KeyError(_A ) if item is _deleted: continue if item.key == key: _SCREAMING_SNAKE_CASE = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self , A ) -> Union[str, Any]: for ind in self._iterate_buckets(_A ): _SCREAMING_SNAKE_CASE = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(_A ) def __len__( self ) -> int: return self._len def __iter__( self ) -> Tuple: yield from (item.key for item in self._buckets if item) def __repr__( self ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = ' ,'.join( f'{item.key}: {item.val}' for item in self._buckets if item ) return f'HashMap({val_string})'
58
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( a , a , a ) -> None: __A : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a ) == len(a ), F"""{len(a )} != {len(a )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) UpperCAmelCase : List[Any] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } UpperCAmelCase : Optional[int] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _SCREAMING_SNAKE_CASE ( a , a ) -> Dict: try: __A : int = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a ) ) def _SCREAMING_SNAKE_CASE ( a , a ) -> List[int]: if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _SCREAMING_SNAKE_CASE ( a , a = "student" , a = None , a = None , a=False , a=None , a=None , **a , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __A : List[str] = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(a , a ): AutoTokenizer.from_pretrained(a ).save_pretrained(a ) # purely for convenience __A : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained(a ).eval() else: assert isinstance(a , a ), F"""teacher must be a model or string got type {type(a )}""" __A : int = teacher.config.to_diff_dict() try: __A , __A : List[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A : str = teacher_e if d is None: __A : List[Any] = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): __A , __A : List[Any] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A : Optional[int] = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A : int = teacher_e if d is None: __A : Optional[Any] = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a ) # Copy weights __A : Dict = teacher.config_class(**a ) __A : int = AutoModelForSeqaSeqLM.from_config(a ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A : Any = student.load_state_dict(teacher.state_dict() , strict=a ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A : Optional[int] = list(range(a ) ), list(range(a ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) if d_layers_to_copy is None: __A : List[int] = pick_layers_to_copy(a , a ) try: if hasattr( a , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a ) copy_layers(teacher.decoder.block , student.decoder.block , a ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __A : Optional[int] = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(a ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
280
0
from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets a_ = '''\ @inproceedings{wang2019glue, title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding}, author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.}, note={In the Proceedings of ICLR.}, year={2019} } ''' a_ = '''\ GLUE, the General Language Understanding Evaluation benchmark (https://gluebenchmark.com/) is a collection of resources for training, evaluating, and analyzing natural language understanding systems. ''' a_ = ''' Compute GLUE evaluation metric associated to each GLUE dataset. Args: predictions: list of predictions to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. Returns: depending on the GLUE subset, one or several of: "accuracy": Accuracy "f1": F1 score "pearson": Pearson Correlation "spearmanr": Spearman Correlation "matthews_correlation": Matthew Correlation Examples: >>> glue_metric = datasets.load_metric(\'glue\', \'sst2\') # \'sst2\' or any of ["mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'mrpc\') # \'mrpc\' or \'qqp\' >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'stsb\') >>> references = [0., 1., 2., 3., 4., 5.] >>> predictions = [0., 1., 2., 3., 4., 5.] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print({"pearson": round(results["pearson"], 2), "spearmanr": round(results["spearmanr"], 2)}) {\'pearson\': 1.0, \'spearmanr\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'cola\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def lowerCamelCase__ ( _a , _a): return float((preds == labels).mean()) def lowerCamelCase__ ( _a , _a): SCREAMING_SNAKE_CASE : Optional[int] = simple_accuracy(_a , _a) SCREAMING_SNAKE_CASE : Tuple = float(fa_score(y_true=_a , y_pred=_a)) return { "accuracy": acc, "f1": fa, } def lowerCamelCase__ ( _a , _a): SCREAMING_SNAKE_CASE : Optional[Any] = float(pearsonr(_a , _a)[0]) SCREAMING_SNAKE_CASE : Tuple = float(spearmanr(_a , _a)[0]) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): '''simple docstring''' def __UpperCamelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), "references": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def __UpperCamelCase ( self : Tuple , a : List[Any] , a : int ) -> int: """simple docstring""" if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(_A , _A )} elif self.config_name == "stsb": return pearson_and_spearman(_A , _A ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(_A , _A ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(_A , _A )} else: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" )
76
def _SCREAMING_SNAKE_CASE ( a , a ) -> list[int]: __A : Optional[int] = int(a ) # Initialize Result __A : Optional[int] = [] # Traverse through all denomination for denomination in reversed(a ): # Find denominations while int(a ) >= int(a ): total_value -= int(a ) answer.append(a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase : List[str] = [] UpperCAmelCase : Optional[int] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): UpperCAmelCase : List[Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(F"""Denomination {i}: """).strip())) UpperCAmelCase : int = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCAmelCase : Tuple = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(F"""Following is minimal change for {value}: """) UpperCAmelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
280
0
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'''vocab_file''': '''vocab.txt''', '''emoji_file''': '''emoji.json'''} UpperCamelCase__ = { '''vocab_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt''', }, '''emoji_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json''', }, } UpperCamelCase__ = { '''abeja/gpt-neox-japanese-2.7b''': 2_0_4_8, } def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' with open(__A, "r", encoding="utf-8" ) as f: UpperCAmelCase__ = json.loads(f.read() ) UpperCAmelCase__ = collections.OrderedDict() UpperCAmelCase__ = collections.OrderedDict() UpperCAmelCase__ = collections.OrderedDict() with open(__A, "r", encoding="utf-8" ) as f: UpperCAmelCase__ = f.readlines() UpperCAmelCase__ = [[t.rstrip("\n" )] if (t == ',' or ',' not in t) else t.rstrip("\n" ).split("," ) for t in token] for idx, b in enumerate(__A ): UpperCAmelCase__ = b UpperCAmelCase__ = idx for wd in b: UpperCAmelCase__ = idx return vocab, raw_vocab, ids_to_tokens, emoji class A ( snake_case__ ): __UpperCAmelCase : Tuple = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : int = ['''input_ids''', '''attention_mask'''] def __init__(self : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : int , __UpperCAmelCase : List[Any]="<|endoftext|>" , __UpperCAmelCase : Union[str, Any]="<|endoftext|>" , __UpperCAmelCase : Union[str, Any]="<|startoftext|>" , __UpperCAmelCase : Dict="<|endoftext|>" , __UpperCAmelCase : Union[str, Any]=False , **__UpperCAmelCase : Dict , ) -> Dict: """simple docstring""" super().__init__( unk_token=_A , pad_token=_A , bos_token=_A , eos_token=_A , do_clean_text=_A , **_A , ) if not os.path.isfile(_A ): raise ValueError( f"""Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained""" " model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) if not os.path.isfile(_A ): raise ValueError( f"""Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google""" " pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) UpperCAmelCase__ = do_clean_text UpperCAmelCase__ = load_vocab_and_emoji(_A , _A ) UpperCAmelCase__ = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def lowercase_ (self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return len(self.raw_vocab ) def lowercase_ (self : str ) -> str: """simple docstring""" return dict(self.raw_vocab , **self.added_tokens_encoder ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : List[str] ) -> Optional[Any]: """simple docstring""" return self.subword_tokenizer.tokenize(_A , clean=self.do_clean_text ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> Tuple: """simple docstring""" return self.vocab.get(_A , self.vocab.get(self.unk_token ) ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return self.subword_tokenizer.convert_id_to_token(_A ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[Any] ) -> str: """simple docstring""" UpperCAmelCase__ = ''.join(_A ).strip() return out_string def lowercase_ (self : str , __UpperCAmelCase : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase__ = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_A , add_special_tokens=_A ) + [self.eos_token_id] ) if len(_A ) > self.model_max_length: UpperCAmelCase__ = input_ids[-self.model_max_length :] return input_ids def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Any = None ) -> Any: """simple docstring""" UpperCAmelCase__ = 0 if os.path.isdir(_A ): UpperCAmelCase__ = os.path.join( _A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( _A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["emoji_file"] ) else: UpperCAmelCase__ = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(_A , "w" , encoding="utf-8" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" " Please check that the vocabulary is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(",".join(_A ) + "\n" ) index += 1 with open(_A , "w" , encoding="utf-8" ) as writer: json.dump(self.emoji , _A ) return vocab_file, emoji_file class A ( snake_case__ ): def __init__(self : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = vocab # same as swe UpperCAmelCase__ = ids_to_tokens # same as bpe UpperCAmelCase__ = emoji UpperCAmelCase__ = np.max([len(_A ) for w in self.vocab.keys()] ) UpperCAmelCase__ = re.compile(r"(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)" ) UpperCAmelCase__ = re.compile(r"[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*" ) UpperCAmelCase__ = re.compile(r"[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}" ) UpperCAmelCase__ = re.compile( r"([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) UpperCAmelCase__ = re.compile( r"(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) UpperCAmelCase__ = re.compile( r"((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*" ) UpperCAmelCase__ = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' UpperCAmelCase__ = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' UpperCAmelCase__ = str.maketrans({k: "<BLOCK>" for k in keisen + blocks} ) def __len__(self : Tuple ) -> Optional[Any]: """simple docstring""" return len(self.ids_to_tokens ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase__ = self.content_repattera.sub("<URL>" , _A ) UpperCAmelCase__ = self.content_repattera.sub("<EMAIL>" , _A ) UpperCAmelCase__ = self.content_repattera.sub("<TEL>" , _A ) UpperCAmelCase__ = self.content_repattera.sub("<DATE>" , _A ) UpperCAmelCase__ = self.content_repattera.sub("<DATE>" , _A ) UpperCAmelCase__ = self.content_repattera.sub("<PRICE>" , _A ) UpperCAmelCase__ = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: UpperCAmelCase__ = content.replace("<BLOCK><BLOCK>" , "<BLOCK>" ) return content def lowercase_ (self : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any]=False ) -> Dict: """simple docstring""" UpperCAmelCase__ = text.replace(" " , "<SP>" ) UpperCAmelCase__ = text.replace(" " , "<SP>" ) UpperCAmelCase__ = text.replace("\r\n" , "<BR>" ) UpperCAmelCase__ = text.replace("\n" , "<BR>" ) UpperCAmelCase__ = text.replace("\r" , "<BR>" ) UpperCAmelCase__ = text.replace("\t" , "<TAB>" ) UpperCAmelCase__ = text.replace("—" , "ー" ) UpperCAmelCase__ = text.replace("−" , "ー" ) for k, v in self.emoji["emoji"].items(): if k in text: UpperCAmelCase__ = text.replace(_A , _A ) if clean: UpperCAmelCase__ = self.clean_text(_A ) def check_simbol(__UpperCAmelCase : int ): UpperCAmelCase__ = x.encode() if len(_A ) == 1 and len(_A ) == 2: UpperCAmelCase__ = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0XC_2A1 and c <= 0XC_2BF) or (c >= 0XC_780 and c <= 0XC_783) or (c >= 0XC_AB9 and c <= 0XC_BBF) or (c >= 0XC_C80 and c <= 0XC_DA2) ): return True return False def checkuae(__UpperCAmelCase : str ): UpperCAmelCase__ = x.encode() if len(_A ) == 1 and len(_A ) == 3: UpperCAmelCase__ = (int(e[0] ) << 1_6) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0XE28_080 and c <= 0XE2B_07F: return True return False UpperCAmelCase__ = 0 UpperCAmelCase__ = [] while pos < len(_A ): UpperCAmelCase__ = min(len(_A ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 UpperCAmelCase__ = [] # (token_id, token, pos) for e in range(_A , _A , -1 ): UpperCAmelCase__ = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(_A ) > 2: UpperCAmelCase__ = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(_A ) > 0: # the smallest token_id is adopted UpperCAmelCase__ = sorted(_A , key=lambda __UpperCAmelCase : x[0] )[0] result.append(_A ) UpperCAmelCase__ = e else: UpperCAmelCase__ = pos + 1 UpperCAmelCase__ = text[pos:end] if check_simbol(_A ): result.append("<KIGOU>" ) elif checkuae(_A ): result.append("<U2000U2BFF>" ) else: for i in wd.encode("utf-8" ): result.append("<|byte%d|>" % i ) UpperCAmelCase__ = end return result def lowercase_ (self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Union[str, Any]="\n" ) -> int: """simple docstring""" UpperCAmelCase__ = [] UpperCAmelCase__ = [] UpperCAmelCase__ = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(_A ) > 0: words.append(bytearray(_A ).decode("utf-8" , errors="replace" ) ) UpperCAmelCase__ = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["emoji_inv"][word] ) elif word == "<SP>": words.append(" " ) elif word == "<BR>": words.append(_A ) elif word == "<TAB>": words.append("\t" ) elif word == "<BLOCK>": words.append("▀" ) elif word == "<KIGOU>": words.append("ǀ" ) elif word == "<U2000U2BFF>": words.append("‖" ) else: words.append(_A ) if len(_A ) > 0: words.append(bytearray(_A ).decode("utf-8" , errors="replace" ) ) UpperCAmelCase__ = ''.join(_A ) return text
65
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _A( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __A : List[Any] = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __A : Union[str, Any] = parent __A : Optional[int] = batch_size __A : int = num_channels __A : int = min_resolution __A : Any = max_resolution __A : List[Any] = do_resize __A : List[Any] = size __A : Union[str, Any] = do_normalize __A : Optional[int] = image_mean __A : Optional[int] = image_std __A : int = do_rescale __A : str = rescale_factor __A : Tuple = do_pad def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase_ ( self , _A , _A=False ): if not batched: __A : List[str] = image_inputs[0] if isinstance(_A , Image.Image ): __A , __A : int = image.size else: __A , __A : Any = image.shape[1], image.shape[2] if w < h: __A : List[Any] = int(self.size['shortest_edge'] * h / w ) __A : List[Any] = self.size['shortest_edge'] elif w > h: __A : Union[str, Any] = self.size['shortest_edge'] __A : str = int(self.size['shortest_edge'] * w / h ) else: __A : Dict = self.size['shortest_edge'] __A : str = self.size['shortest_edge'] else: __A : int = [] for image in image_inputs: __A , __A : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __A : List[str] = max(_A , key=lambda _A : item[0] )[0] __A : str = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase : List[str] = YolosImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): __A : Dict = YolosImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): __A : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) def UpperCAmelCase_ ( self ): __A : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , _A ) __A : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _A ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing __A : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input __A : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A , __A : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) __A : str = image_processing(_A , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input __A : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : List[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Tuple = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing __A : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input __A : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __A , __A : Union[str, Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __A : Optional[int] = image_processing(_A , return_tensors='pt' ).pixel_values __A , __A : Optional[int] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processings __A : Tuple = self.image_processing_class(**self.image_processor_dict ) __A : Any = self.image_processing_class(do_resize=_A , do_normalize=_A , do_rescale=_A ) # create random PyTorch tensors __A : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __A : Optional[int] = image_processing_a.pad(_A , return_tensors='pt' ) __A : Optional[int] = image_processing_a(_A , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1e-4 ) ) @slow def UpperCAmelCase_ ( self ): # prepare image and target __A : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __A : Optional[Any] = json.loads(f.read() ) __A : Optional[Any] = {'image_id': 39769, 'annotations': target} # encode them __A : str = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) __A : List[Any] = image_processing(images=_A , annotations=_A , return_tensors='pt' ) # verify pixel values __A : List[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : List[Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Any = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify orig_size __A : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : str = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) ) @slow def UpperCAmelCase_ ( self ): # prepare image, target and masks_path __A : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __A : Tuple = json.loads(f.read() ) __A : Any = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __A : List[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __A : Any = YolosImageProcessor(format='coco_panoptic' ) __A : List[Any] = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='pt' ) # verify pixel values __A : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , _A ) __A : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _A , atol=1e-4 ) ) # verify area __A : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _A ) ) # verify boxes __A : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _A ) __A : Optional[Any] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _A , atol=1e-3 ) ) # verify image_id __A : Union[str, Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _A ) ) # verify is_crowd __A : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _A ) ) # verify class_labels __A : List[str] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _A ) ) # verify masks __A : Tuple = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _A ) # verify orig_size __A : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _A ) ) # verify size __A : int = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _A ) )
280
0
"""simple docstring""" import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument lowercase__ = { '''/attention/''': '''/0/SelfAttention/''', '''/self_attention/''': '''/0/SelfAttention/''', '''/encoder_decoder_attention/''': '''/1/EncDecAttention/''', '''value''': '''v''', '''query''': '''q''', '''key''': '''k''', '''out''': '''o''', '''pre_self_attention_layer_norm''': '''0/layer_norm''', '''pre_cross_attention_layer_norm''': '''1/layer_norm''', '''pre_attention_layer_norm''': '''0/layer_norm''', # previously 1, but seems wrong '''token_embedder''': '''shared''', '''encoder_norm''': '''final_layer_norm''', '''decoder_norm''': '''final_layer_norm''', '''relpos_bias/rel_embedding''': '''block/0/layer/0/SelfAttention/relative_attention_bias/weight''', '''router/router_weights/w/''': '''router/classifier/''', '''roer/roer_weights/w/''': '''router/classifier/''', '''logits_dense''': '''lm_head''', } def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[Any]: # 1. in HF T5, we have block.{x}.layer.{y}. which corresponds to layer.{x} in # the original model a__: int = list(s_dict.keys() ) for key in keys: a__: Union[str, Any] = r'.*/layers_(\d+)' a__: Dict = key if re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = re.sub(r'layers_(\d+)' , r'block/\1/layer' , _SCREAMING_SNAKE_CASE ) a__: List[Any] = r'(encoder|decoder)\/' if re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Tuple = re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).groups() if groups[0] == "encoder": a__: Union[str, Any] = re.sub(r'/mlp/' , r'/1/mlp/' , _SCREAMING_SNAKE_CASE ) a__: Any = re.sub(r'/pre_mlp_layer_norm/' , r'/1/layer_norm/' , _SCREAMING_SNAKE_CASE ) elif groups[0] == "decoder": a__: Optional[int] = re.sub(r'/mlp/' , r'/2/mlp/' , _SCREAMING_SNAKE_CASE ) a__: Optional[int] = re.sub(r'/pre_mlp_layer_norm/' , r'/2/layer_norm/' , _SCREAMING_SNAKE_CASE ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: a__: str = new_key.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(F'{key} -> {new_key}' ) a__: Dict = s_dict.pop(_SCREAMING_SNAKE_CASE ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: a__: Any = s_dict[ 'encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight' ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: a__: List[Any] = s_dict[ 'decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight' ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: a__: Optional[int] = s_dict[key].shape[0] a__: Tuple = s_dict[key] for idx in range(_SCREAMING_SNAKE_CASE ): a__: str = expert_weihts[idx] print(F'{key} -> {key.replace("expert/" , "nested fstring" )}' ) s_dict.pop(_SCREAMING_SNAKE_CASE ) return s_dict lowercase__ = { '''NUM_ENCODER_LAYERS''': '''num_layers''', '''NUM_DECODER_LAYERS''': '''num_decoder_layers''', '''NUM_HEADS''': '''num_heads''', '''HEAD_DIM''': '''d_kv''', '''EMBED_DIM''': '''d_model''', '''MLP_DIM''': '''d_ff''', '''NUM_SELECTED_EXPERTS''': '''num_selected_experts''', '''NUM_ENCODER_SPARSE_LAYERS''': '''num_sparse_encoder_layers''', '''NUM_DECODER_SPARSE_LAYERS''': '''num_sparse_decoder_layers''', '''dense.MlpBlock.activations''': '''feed_forward_proj''', } def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: # Convert a google style config to the hugging face fromat import regex as re with open(_SCREAMING_SNAKE_CASE , 'r' ) as f: a__: List[str] = f.read() a__: int = re.findall(r'(.*) = ([0-9.]*)' , _SCREAMING_SNAKE_CASE ) a__: Optional[Any] = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": a__: str = float(_SCREAMING_SNAKE_CASE ) if '.' in value else int(_SCREAMING_SNAKE_CASE ) a__: Any = re.findall(r'(.*activations) = \(\'(.*)\',\)' , _SCREAMING_SNAKE_CASE )[0] a__: List[Any] = str(activation[1] ) a__: List[str] = num_experts a__: Union[str, Any] = SwitchTransformersConfig(**_SCREAMING_SNAKE_CASE ) return config def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE="./" , _SCREAMING_SNAKE_CASE=8 ) ->List[Any]: # Initialise PyTorch model print(F'Loading flax weights from : {flax_checkpoint_path}' ) a__: List[Any] = checkpoints.load_tax_checkpoint(_SCREAMING_SNAKE_CASE ) if gin_file is not None: a__: str = convert_gin_to_config(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: a__: int = SwitchTransformersConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = SwitchTransformersForConditionalGeneration(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = flax_params['target'] a__: Optional[int] = flatten_dict(_SCREAMING_SNAKE_CASE , sep='/' ) a__: Union[str, Any] = rename_keys(_SCREAMING_SNAKE_CASE ) a__: Any = unflatten_dict(_SCREAMING_SNAKE_CASE , sep='/' ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(F'Save PyTorch model to {pytorch_dump_path}' ) pt_model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the' ' model architecture. If not provided, a `gin_file` has to be provided.' ), ) parser.add_argument( '--gin_file', default=None, type=str, required=False, help='Path to the gin config file. If not provided, a `config_file` has to be passed ', ) parser.add_argument( '--config_name', default=None, type=str, required=False, help='Config name of SwitchTransformers model.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output pytorch model.' ) parser.add_argument('--num_experts', default=8, type=int, required=False, help='Number of experts') lowercase__ = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
290
import argparse import json from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( ) -> List[Any]: __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=a , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=a , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=a , help='where to store parsed gold_data_path file' , ) __A : Optional[int] = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: __A : List[Any] = json.load(a ) for dpr_record in tqdm(a ): __A : Dict = dpr_record['question'] __A : Any = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(a ) + '\n' ) if __name__ == "__main__": main()
280
0
class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : Optional[int] = None lowerCamelCase : List[Any] = None lowerCamelCase : Tuple = graph self._normalize_graph(_A , _A ) lowerCamelCase : str = len(_A ) lowerCamelCase : str = None def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: if sources is int: lowerCamelCase : int = [sources] if sinks is int: lowerCamelCase : Any = [sinks] if len(_A ) == 0 or len(_A ) == 0: return lowerCamelCase : List[Any] = sources[0] lowerCamelCase : Tuple = sinks[0] # make fake vertex if there are more # than one source or sink if len(_A ) > 1 or len(_A ) > 1: lowerCamelCase : Optional[int] = 0 for i in sources: max_input_flow += sum(self.graph[i] ) lowerCamelCase : List[str] = len(self.graph ) + 1 for room in self.graph: room.insert(0 , 0 ) self.graph.insert(0 , [0] * size ) for i in sources: lowerCamelCase : Any = max_input_flow lowerCamelCase : Union[str, Any] = 0 lowerCamelCase : Optional[int] = len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: lowerCamelCase : Dict = max_input_flow lowerCamelCase : Optional[Any] = size - 1 def _lowercase ( self ) -> Any: if self.maximum_flow_algorithm is None: raise Exception("You need to set maximum flow algorithm before." ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def _lowercase ( self , UpperCamelCase__ ) -> Tuple: lowerCamelCase : Any = algorithm(self ) class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ ) -> List[str]: lowerCamelCase : str = flow_network lowerCamelCase : int = flow_network.verticesCount lowerCamelCase : Any = flow_network.sourceIndex lowerCamelCase : Tuple = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that lowerCamelCase : Tuple = flow_network.graph lowerCamelCase : List[str] = False def _lowercase ( self ) -> Tuple: if not self.executed: self._algorithm() lowerCamelCase : Any = True def _lowercase ( self ) -> Optional[Any]: pass class UpperCamelCase__ (snake_case__ ): '''simple docstring''' def __init__( self , UpperCamelCase__ ) -> str: super().__init__(_A ) # use this to save your result lowerCamelCase : List[Any] = -1 def _lowercase ( self ) -> Optional[Any]: if not self.executed: raise Exception("You should execute algorithm before using its result!" ) return self.maximum_flow class UpperCamelCase__ (snake_case__ ): '''simple docstring''' def __init__( self , UpperCamelCase__ ) -> int: super().__init__(_A ) lowerCamelCase : Any = [[0] * self.verticies_count for i in range(self.verticies_count )] lowerCamelCase : str = [0] * self.verticies_count lowerCamelCase : List[str] = [0] * self.verticies_count def _lowercase ( self ) -> Tuple: lowerCamelCase : Any = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule lowerCamelCase : Optional[Any] = [ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list lowerCamelCase : Union[str, Any] = 0 while i < len(_A ): lowerCamelCase : int = vertices_list[i] lowerCamelCase : Dict = self.heights[vertex_index] self.process_vertex(_A ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(_A ) ) lowerCamelCase : int = 0 else: i += 1 lowerCamelCase : Optional[Any] = sum(self.preflow[self.source_index] ) def _lowercase ( self , UpperCamelCase__ ) -> List[str]: while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_A , _A ) self.relabel(_A ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Tuple: lowerCamelCase : Union[str, Any] = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def _lowercase ( self , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : List[Any] = None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): lowerCamelCase : Union[str, Any] = self.heights[to_index] if min_height is not None: lowerCamelCase : List[str] = min_height + 1 if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : int = [0] SCREAMING_SNAKE_CASE__ : Any = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] SCREAMING_SNAKE_CASE__ : List[str] = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network SCREAMING_SNAKE_CASE__ : int = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate SCREAMING_SNAKE_CASE__ : Optional[Any] = flow_network.find_maximum_flow() print(f'''maximum flow is {maximum_flow}''')
48
from heapq import heappop, heappush import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a , ) -> tuple[float | int, list[tuple[int, int]]]: __A , __A : int = grid.shape __A : Any = [-1, 1, 0, 0] __A : Optional[Any] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __A , __A : Optional[int] = [(0, source)], set() __A : Any = np.full((rows, cols) , np.inf ) __A : Any = 0 __A : Any = np.empty((rows, cols) , dtype=a ) __A : Optional[Any] = None while queue: ((__A) , (__A)) : List[str] = heappop(a ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __A : int = [] while (x, y) != source: path.append((x, y) ) __A , __A : Optional[int] = predecessors[x, y] path.append(a ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a ) ): __A , __A : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __A : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a , (dist + 1, (nx, ny)) ) __A : List[Any] = dist + 1 __A : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
280
0
import re import string import numpy as np import datasets lowercase = ''' Returns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list. ''' lowercase = ''' Args: predictions: List of predicted texts. references: List of reference texts. regexes_to_ignore: List, defaults to None. Regex expressions of characters to ignore when calculating the exact matches. Note: these regexes are removed from the input data before the changes based on the options below (e.g. ignore_case, ignore_punctuation, ignore_numbers) are applied. ignore_case: Boolean, defaults to False. If true, turns everything to lowercase so that capitalization differences are ignored. ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. Returns: exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive. Examples: >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results["exact_match"], 1)) 25.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True) >>> print(round(results["exact_match"], 1)) 50.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True) >>> print(round(results["exact_match"], 1)) 75.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True) >>> print(round(results["exact_match"], 1)) 100.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."] >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results["exact_match"], 1)) 33.3 ''' lowercase = ''' ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase_ ( datasets.Metric ): '''simple docstring''' def _UpperCamelCase ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , reference_urls=[] , ) def _UpperCamelCase ( self , a , a , a=None , a=False , a=False , a=False , ) -> Optional[int]: if regexes_to_ignore is not None: for s in regexes_to_ignore: snake_case_ = np.array([re.sub(_A , '' , _A ) for x in predictions] ) snake_case_ = np.array([re.sub(_A , '' , _A ) for x in references] ) else: snake_case_ = np.asarray(_A ) snake_case_ = np.asarray(_A ) if ignore_case: snake_case_ = np.char.lower(_A ) snake_case_ = np.char.lower(_A ) if ignore_punctuation: snake_case_ = string.punctuation.maketrans('' , '' , string.punctuation ) snake_case_ = np.char.translate(_A , table=_A ) snake_case_ = np.char.translate(_A , table=_A ) if ignore_numbers: snake_case_ = string.digits.maketrans('' , '' , string.digits ) snake_case_ = np.char.translate(_A , table=_A ) snake_case_ = np.char.translate(_A , table=_A ) snake_case_ = predictions == references return {"exact_match": np.mean(_A ) * 1_00}
178
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
0
'''simple docstring''' class a_ : def __init__( self , snake_case_ ): _lowerCAmelCase : Tuple = set_counts _lowerCAmelCase : Union[str, Any] = max(_A ) _lowerCAmelCase : List[str] = len(_A ) _lowerCAmelCase : Union[str, Any] = [1] * num_sets _lowerCAmelCase : int = list(range(_A ) ) def __UpperCamelCase ( self , snake_case_ , snake_case_ ): _lowerCAmelCase : Tuple = self.get_parent(_A ) _lowerCAmelCase : Dict = self.get_parent(_A ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : int = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 _lowerCAmelCase : List[str] = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] _lowerCAmelCase : Any = 0 _lowerCAmelCase : Optional[int] = src_parent _lowerCAmelCase : Optional[Any] = self.set_counts[src_parent] _lowerCAmelCase : List[str] = max(self.max_set , _A ) return True def __UpperCamelCase ( self , snake_case_ ): if self.parents[disj_set] == disj_set: return disj_set _lowerCAmelCase : Union[str, Any] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
309
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() UpperCAmelCase : List[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _SCREAMING_SNAKE_CASE ( a , a ) -> Optional[Any]: __A : Any = { 'attention_cell': 'multi_head', 'num_layers': 4, 'units': 10_24, 'hidden_size': 7_68, 'max_length': 5_12, 'num_heads': 8, 'scaled': True, 'dropout': 0.1, 'use_residual': True, 'embed_size': 10_24, 'embed_dropout': 0.1, 'word_embed': None, 'layer_norm_eps': 1e-5, 'token_type_vocab_size': 2, } __A : str = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __A : Optional[int] = BERTEncoder( attention_cell=predefined_args['attention_cell'] , num_layers=predefined_args['num_layers'] , units=predefined_args['units'] , hidden_size=predefined_args['hidden_size'] , max_length=predefined_args['max_length'] , num_heads=predefined_args['num_heads'] , scaled=predefined_args['scaled'] , dropout=predefined_args['dropout'] , output_attention=a , output_all_encodings=a , use_residual=predefined_args['use_residual'] , activation=predefined_args.get('activation' , 'gelu' ) , layer_norm_eps=predefined_args.get('layer_norm_eps' , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __A : Union[str, Any] = 'openwebtext_ccnews_stories_books_cased' # Specify download folder to Gluonnlp's vocab __A : Any = os.path.join(get_home_dir() , 'models' ) __A : List[Any] = _load_vocab(a , a , a , cls=a ) __A : Dict = nlp.model.BERTModel( a , len(a ) , units=predefined_args['units'] , embed_size=predefined_args['embed_size'] , embed_dropout=predefined_args['embed_dropout'] , word_embed=predefined_args['word_embed'] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args['token_type_vocab_size'] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) __A : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __A : Any = { 'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': predefined_args['dropout'], 'hidden_act': 'gelu', 'hidden_dropout_prob': predefined_args['dropout'], 'hidden_size': predefined_args['embed_size'], 'initializer_range': 0.02, 'intermediate_size': predefined_args['hidden_size'], 'layer_norm_eps': predefined_args['layer_norm_eps'], 'max_position_embeddings': predefined_args['max_length'], 'model_type': 'bort', 'num_attention_heads': predefined_args['num_heads'], 'num_hidden_layers': predefined_args['num_layers'], 'pad_token_id': 1, # 2 = BERT, 1 = RoBERTa 'type_vocab_size': 1, # 2 = BERT, 1 = RoBERTa 'vocab_size': len(a ), } __A : int = BertConfig.from_dict(a ) __A : Union[str, Any] = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a , a ): __A : Tuple = hf_param.shape __A : str = to_torch(params[gluon_param] ) __A : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), F"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __A : str = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , 'word_embed.0.weight' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , 'encoder.position_weight' ) __A : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , 'encoder.layer_norm.beta' ) __A : Tuple = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , 'encoder.layer_norm.gamma' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __A : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __A : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __A : BertSelfAttention = layer.attention.self __A : Optional[Any] = check_and_map_params( self_attn.key.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.key.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.query.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __A : Optional[Any] = check_and_map_params( self_attn.query.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __A : Union[str, Any] = check_and_map_params( self_attn.value.bias.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __A : Optional[int] = check_and_map_params( self_attn.value.weight.data , F"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __A : BertSelfOutput = layer.attention.output __A : Tuple = check_and_map_params( self_output.dense.bias , F"""encoder.transformer_cells.{i}.proj.bias""" ) __A : int = check_and_map_params( self_output.dense.weight , F"""encoder.transformer_cells.{i}.proj.weight""" ) __A : List[Any] = check_and_map_params( self_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __A : str = check_and_map_params( self_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __A : BertIntermediate = layer.intermediate __A : int = check_and_map_params( intermediate.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __A : List[Any] = check_and_map_params( intermediate.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __A : BertOutput = layer.output __A : List[Any] = check_and_map_params( bert_output.dense.bias , F"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __A : Dict = check_and_map_params( bert_output.dense.weight , F"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __A : Optional[int] = check_and_map_params( bert_output.LayerNorm.bias , F"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __A : Dict = check_and_map_params( bert_output.LayerNorm.weight , F"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __A : Any = RobertaTokenizer.from_pretrained('roberta-base' ) __A : List[str] = tokenizer.encode_plus(a )['input_ids'] # Get gluon output __A : List[str] = mx.nd.array([input_ids] ) __A : Union[str, Any] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) __A : Optional[Any] = BertModel.from_pretrained(a ) hf_bort_model.eval() __A : Tuple = tokenizer.encode_plus(a , return_tensors='pt' ) __A : Any = hf_bort_model(**a )[0] __A : Union[str, Any] = output_gluon[0].asnumpy() __A : Tuple = output_hf[0].detach().numpy() __A : int = np.max(np.abs(hf_layer - gluon_layer ) ).item() __A : int = np.allclose(a , a , atol=1e-3 ) if success: print('✔️ Both model do output the same tensors' ) else: print('❌ Both model do **NOT** output the same tensors' ) print('Absolute difference is:' , a ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase : Dict = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
280
0
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowercase_ = 1_6 lowercase_ = 3_2 def a ( A__ : int , A__ : Optional[Any] = 16 ) -> Tuple: """simple docstring""" _lowercase =AutoTokenizer.from_pretrained('bert-base-cased' ) _lowercase =load_dataset('glue' , 'mrpc' ) def tokenize_function(A__ : Dict ): # max_length=None => use the model max length (it's actually the default) _lowercase =tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=A__ , max_length=A__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _lowercase =datasets.map( A__ , batched=A__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowercase =tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(A__ : str ): # On TPU it's best to pad everything to the same length or training will be very slow. _lowercase =128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _lowercase =16 elif accelerator.mixed_precision != "no": _lowercase =8 else: _lowercase =None return tokenizer.pad( A__ , padding='longest' , max_length=A__ , pad_to_multiple_of=A__ , return_tensors='pt' , ) # Instantiate dataloaders. _lowercase =DataLoader( tokenized_datasets['train'] , shuffle=A__ , collate_fn=A__ , batch_size=A__ , drop_last=A__ ) _lowercase =DataLoader( tokenized_datasets['validation'] , shuffle=A__ , collate_fn=A__ , batch_size=A__ , drop_last=(accelerator.mixed_precision == 'fp8') , ) return train_dataloader, eval_dataloader def a ( A__ : int , A__ : str ) -> int: """simple docstring""" _lowercase =Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowercase =config['lr'] _lowercase =int(config['num_epochs'] ) _lowercase =int(config['seed'] ) _lowercase =int(config['batch_size'] ) _lowercase =evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _lowercase =1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _lowercase =batch_size // MAX_GPU_BATCH_SIZE _lowercase =MAX_GPU_BATCH_SIZE set_seed(A__ ) _lowercase =get_dataloaders(A__ , A__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowercase =AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=A__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _lowercase =model.to(accelerator.device ) # Instantiate optimizer _lowercase =AdamW(params=model.parameters() , lr=A__ ) # Instantiate scheduler _lowercase =get_linear_schedule_with_warmup( optimizer=A__ , num_warmup_steps=100 , num_training_steps=(len(A__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowercase =accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # Now we train the model for epoch in range(A__ ): model.train() for step, batch in enumerate(A__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _lowercase =model(**A__ ) _lowercase =outputs.loss _lowercase =loss / gradient_accumulation_steps accelerator.backward(A__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(A__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowercase =model(**A__ ) _lowercase =outputs.logits.argmax(dim=-1 ) _lowercase =accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=A__ , references=A__ , ) _lowercase =metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , A__ ) def a ( ) -> Dict: """simple docstring""" _lowercase =argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=A__ , default=A__ , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) _lowercase =parser.parse_args() _lowercase ={'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(A__ , A__ ) if __name__ == "__main__": main()
205
import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( a , a , a ) -> float: __A : List[str] = x __A : str = y for step in range(a ): # noqa: B007 __A : Union[str, Any] = a * a - b * b + x __A : Optional[int] = 2 * a * b + y __A : List[str] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _SCREAMING_SNAKE_CASE ( a ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( a = 8_00 , a = 6_00 , a = -0.6 , a = 0 , a = 3.2 , a = 50 , a = True , ) -> Image.Image: __A : str = Image.new('RGB' , (image_width, image_height) ) __A : Dict = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates __A : Dict = figure_width / image_width * image_height __A : Union[str, Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width __A : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height __A : Union[str, Any] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __A : Optional[Any] = get_color_coded_rgb(a ) else: __A : Dict = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCAmelCase : str = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
280
0