code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" __lowercase = { """Pillow""": """Pillow<10.0.0""", """accelerate""": """accelerate>=0.20.3""", """av""": """av==9.2.0""", """beautifulsoup4""": """beautifulsoup4""", """black""": """black~=23.1""", """codecarbon""": """codecarbon==1.2.0""", """cookiecutter""": """cookiecutter==1.7.3""", """dataclasses""": """dataclasses""", """datasets""": """datasets!=2.5.0""", """decord""": """decord==0.6.0""", """deepspeed""": """deepspeed>=0.9.3""", """diffusers""": """diffusers""", """dill""": """dill<0.3.5""", """evaluate""": """evaluate>=0.2.0""", """fairscale""": """fairscale>0.3""", """faiss-cpu""": """faiss-cpu""", """fastapi""": """fastapi""", """filelock""": """filelock""", """flax""": """flax>=0.4.1,<=0.7.0""", """ftfy""": """ftfy""", """fugashi""": """fugashi>=1.0""", """GitPython""": """GitPython<3.1.19""", """hf-doc-builder""": """hf-doc-builder>=0.3.0""", """huggingface-hub""": """huggingface-hub>=0.14.1,<1.0""", """importlib_metadata""": """importlib_metadata""", """ipadic""": """ipadic>=1.0.0,<2.0""", """isort""": """isort>=5.5.4""", """jax""": """jax>=0.2.8,!=0.3.2,<=0.4.13""", """jaxlib""": """jaxlib>=0.1.65,<=0.4.13""", """jieba""": """jieba""", """kenlm""": """kenlm""", """keras-nlp""": """keras-nlp>=0.3.1""", """librosa""": """librosa""", """nltk""": """nltk""", """natten""": """natten>=0.14.6""", """numpy""": """numpy>=1.17""", """onnxconverter-common""": """onnxconverter-common""", """onnxruntime-tools""": """onnxruntime-tools>=1.4.2""", """onnxruntime""": """onnxruntime>=1.4.0""", """opencv-python""": """opencv-python""", """optuna""": """optuna""", """optax""": """optax>=0.0.8,<=0.1.4""", """packaging""": """packaging>=20.0""", """parameterized""": """parameterized""", """phonemizer""": """phonemizer""", """protobuf""": """protobuf""", """psutil""": """psutil""", """pyyaml""": """pyyaml>=5.1""", """pydantic""": """pydantic<2""", """pytest""": """pytest>=7.2.0""", """pytest-timeout""": """pytest-timeout""", """pytest-xdist""": """pytest-xdist""", """python""": """python>=3.8.0""", """ray[tune]""": """ray[tune]""", """regex""": """regex!=2019.12.17""", """requests""": """requests""", """rhoknp""": """rhoknp>=1.1.0,<1.3.1""", """rjieba""": """rjieba""", """rouge-score""": """rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1""", """ruff""": """ruff>=0.0.241,<=0.0.259""", """sacrebleu""": """sacrebleu>=1.4.12,<2.0.0""", """sacremoses""": """sacremoses""", """safetensors""": """safetensors>=0.3.1""", """sagemaker""": """sagemaker>=2.31.0""", """scikit-learn""": """scikit-learn""", """sentencepiece""": """sentencepiece>=0.1.91,!=0.1.92""", """sigopt""": """sigopt""", """starlette""": """starlette""", """sudachipy""": """sudachipy>=0.6.6""", """sudachidict_core""": """sudachidict_core>=20220729""", """tensorflow-cpu""": """tensorflow-cpu>=2.6,<2.14""", """tensorflow""": """tensorflow>=2.6,<2.14""", """tensorflow-text""": """tensorflow-text<2.14""", """tf2onnx""": """tf2onnx""", """timeout-decorator""": """timeout-decorator""", """timm""": """timm""", """tokenizers""": """tokenizers>=0.11.1,!=0.11.3,<0.14""", """torch""": """torch>=1.9,!=1.12.0""", """torchaudio""": """torchaudio""", """torchvision""": """torchvision""", """pyctcdecode""": """pyctcdecode>=0.4.0""", """tqdm""": """tqdm>=4.27""", """unidic""": """unidic>=1.0.2""", """unidic_lite""": """unidic_lite>=1.0.7""", """urllib3""": """urllib3<2.0.0""", """uvicorn""": """uvicorn""", }
40
'''simple docstring''' import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger a__ : Any = get_logger(__name__) class UpperCAmelCase__ : def __init__( self , lowercase = None ) -> List[str]: __UpperCamelCase = ( os.path.join(lowercase , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) __UpperCamelCase = Extractor def __lowerCamelCase ( self , lowercase ) -> str: from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" __UpperCamelCase = os.path.abspath(lowercase ) return os.path.join(self.extract_dir , hash_url_to_filename(lowercase ) ) def __lowerCamelCase ( self , lowercase , lowercase ) -> bool: return force_extract or ( not os.path.isfile(lowercase ) and not (os.path.isdir(lowercase ) and os.listdir(lowercase )) ) def __lowerCamelCase ( self , lowercase , lowercase = False ) -> str: __UpperCamelCase = self.extractor.infer_extractor_format(lowercase ) if not extractor_format: return input_path __UpperCamelCase = self._get_output_path(lowercase ) if self._do_extract(lowercase , lowercase ): self.extractor.extract(lowercase , lowercase , lowercase ) return output_path class UpperCAmelCase__ ( UpperCAmelCase_): @classmethod @abstractmethod def __lowerCamelCase ( cls , lowercase , **lowercase ) -> bool: ... @staticmethod @abstractmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: ... class UpperCAmelCase__ ( UpperCAmelCase_ , UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> int: with open(lowercase , """rb""" ) as f: return f.read(lowercase ) @classmethod def __lowerCamelCase ( cls , lowercase , lowercase = b"" ) -> bool: if not magic_number: __UpperCamelCase = max(len(lowercase ) for cls_magic_number in cls.magic_numbers ) try: __UpperCamelCase = cls.read_magic_number(lowercase , lowercase ) except OSError: return False return any(magic_number.startswith(lowercase ) for cls_magic_number in cls.magic_numbers ) class UpperCAmelCase__ ( UpperCAmelCase_): @classmethod def __lowerCamelCase ( cls , lowercase , **lowercase ) -> bool: return tarfile.is_tarfile(lowercase ) @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> str: def resolved(lowercase ) -> str: return os.path.realpath(os.path.abspath(lowercase ) ) def badpath(lowercase , lowercase ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(lowercase , lowercase ) ).startswith(lowercase ) def badlink(lowercase , lowercase ) -> bool: # Links are interpreted relative to the directory containing the link __UpperCamelCase = resolved(os.path.join(lowercase , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=lowercase ) __UpperCamelCase = resolved(lowercase ) for finfo in members: if badpath(finfo.name , lowercase ): logger.error(f"Extraction of {finfo.name} is blocked (illegal path)" ) elif finfo.issym() and badlink(lowercase , lowercase ): logger.error(f"Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}" ) elif finfo.islnk() and badlink(lowercase , lowercase ): logger.error(f"Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}" ) else: yield finfo @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: os.makedirs(lowercase , exist_ok=lowercase ) __UpperCamelCase = tarfile.open(lowercase ) tar_file.extractall(lowercase , members=TarExtractor.safemembers(lowercase , lowercase ) ) tar_file.close() class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\x1F\x8B'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: with gzip.open(lowercase , """rb""" ) as gzip_file: with open(lowercase , """wb""" ) as extracted_file: shutil.copyfileobj(lowercase , lowercase ) class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [ B'''PK\x03\x04''', B'''PK\x05\x06''', # empty archive B'''PK\x07\x08''', # spanned archive ] @classmethod def __lowerCamelCase ( cls , lowercase , lowercase = b"" ) -> bool: if super().is_extractable(lowercase , magic_number=lowercase ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(lowercase , """rb""" ) as fp: __UpperCamelCase = _EndRecData(lowercase ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: __UpperCamelCase = fp.read(lowercase ) # CD is where we expect it to be if len(lowercase ) == sizeCentralDir: __UpperCamelCase = struct.unpack(lowercase , lowercase ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: os.makedirs(lowercase , exist_ok=lowercase ) with zipfile.ZipFile(lowercase , """r""" ) as zip_file: zip_file.extractall(lowercase ) zip_file.close() class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\xFD\x37\x7A\x58\x5A\x00'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: with lzma.open(lowercase ) as compressed_file: with open(lowercase , """wb""" ) as extracted_file: shutil.copyfileobj(lowercase , lowercase ) class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''Rar!\x1a\x07\x00''', B'''Rar!\x1a\x07\x01\x00'''] # RAR_ID # RAR5_ID @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(lowercase , exist_ok=lowercase ) __UpperCamelCase = rarfile.RarFile(lowercase ) rf.extractall(lowercase ) rf.close() class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\x28\xb5\x2F\xFD'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd __UpperCamelCase = zstd.ZstdDecompressor() with open(lowercase , """rb""" ) as ifh, open(lowercase , """wb""" ) as ofh: dctx.copy_stream(lowercase , lowercase ) class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\x42\x5A\x68'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: with bza.open(lowercase , """rb""" ) as compressed_file: with open(lowercase , """wb""" ) as extracted_file: shutil.copyfileobj(lowercase , lowercase ) class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\x37\x7A\xBC\xAF\x27\x1C'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(lowercase , exist_ok=lowercase ) with pyazr.SevenZipFile(lowercase , """r""" ) as archive: archive.extractall(lowercase ) class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = [B'''\x04\x22\x4D\x18'''] @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> None: if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(lowercase , """rb""" ) as compressed_file: with open(lowercase , """wb""" ) as extracted_file: shutil.copyfileobj(lowercase , lowercase ) class UpperCAmelCase__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) __SCREAMING_SNAKE_CASE = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def __lowerCamelCase ( cls ) -> Union[str, Any]: return max( len(lowercase ) for extractor in cls.extractors.values() if issubclass(lowercase , lowercase ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def __lowerCamelCase ( lowercase , lowercase ) -> str: try: return MagicNumberBaseExtractor.read_magic_number(lowercase , magic_number_length=lowercase ) except OSError: return b"" @classmethod def __lowerCamelCase ( cls , lowercase , lowercase = False ) -> bool: warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=lowercase , ) __UpperCamelCase = cls.infer_extractor_format(lowercase ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def __lowerCamelCase ( cls , lowercase ) -> str: # <Added version="2.4.0"/> __UpperCamelCase = cls._get_magic_number_max_length() __UpperCamelCase = cls._read_magic_number(lowercase , lowercase ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(lowercase , magic_number=lowercase ): return extractor_format @classmethod def __lowerCamelCase ( cls , lowercase , lowercase , lowercase = None , lowercase = "deprecated" , ) -> None: os.makedirs(os.path.dirname(lowercase ) , exist_ok=lowercase ) # Prevent parallel extractions __UpperCamelCase = str(Path(lowercase ).with_suffix(""".lock""" ) ) with FileLock(lowercase ): shutil.rmtree(lowercase , ignore_errors=lowercase ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(lowercase , lowercase ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=lowercase , ) __UpperCamelCase = extractor if extractor != """deprecated""" else extractor_format else: __UpperCamelCase = cls.extractors[extractor_format] return extractor.extract(lowercase , lowercase ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=lowercase , ) for extractor in cls.extractors.values(): if extractor.is_extractable(lowercase ): return extractor.extract(lowercase , lowercase )
349
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowercase = { """configuration_owlvit""": [ """OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """OwlViTConfig""", """OwlViTOnnxConfig""", """OwlViTTextConfig""", """OwlViTVisionConfig""", ], """processing_owlvit""": ["""OwlViTProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase = ["""OwlViTFeatureExtractor"""] lowercase = ["""OwlViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase = [ """OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """OwlViTModel""", """OwlViTPreTrainedModel""", """OwlViTTextModel""", """OwlViTVisionModel""", """OwlViTForObjectDetection""", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys lowercase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
364
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase = logging.get_logger(__name__) lowercase = { """kakaobrain/align-base""": """https://huggingface.co/kakaobrain/align-base/resolve/main/config.json""", } class __lowercase ( A ): '''simple docstring''' _A : int = '''align_text_model''' def __init__( self : Tuple , _a : Tuple=30_522 , _a : str=768 , _a : Tuple=12 , _a : Dict=12 , _a : Any=3_072 , _a : str="gelu" , _a : int=0.1 , _a : Optional[Any]=0.1 , _a : int=512 , _a : List[str]=2 , _a : Any=0.02 , _a : Dict=1E-12 , _a : Tuple=0 , _a : Optional[Any]="absolute" , _a : str=True , **_a : Union[str, Any] , ): super().__init__(**_a ) UpperCamelCase__ = vocab_size UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = hidden_act UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = max_position_embeddings UpperCamelCase__ = type_vocab_size UpperCamelCase__ = initializer_range UpperCamelCase__ = layer_norm_eps UpperCamelCase__ = position_embedding_type UpperCamelCase__ = use_cache UpperCamelCase__ = pad_token_id @classmethod def A_ ( cls : List[str] , _a : Union[str, os.PathLike] , **_a : Any ): cls._set_token_in_kwargs(_a ) UpperCamelCase__ , UpperCamelCase__ = cls.get_config_dict(_a , **_a ) # get the text config dict if we are loading from AlignConfig if config_dict.get('''model_type''' ) == "align": UpperCamelCase__ = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(_a , **_a ) class __lowercase ( A ): '''simple docstring''' _A : List[Any] = '''align_vision_model''' def __init__( self : List[str] , _a : int = 3 , _a : int = 600 , _a : float = 2.0 , _a : float = 3.1 , _a : int = 8 , _a : List[int] = [3, 3, 5, 3, 5, 5, 3] , _a : List[int] = [32, 16, 24, 40, 80, 112, 192] , _a : List[int] = [16, 24, 40, 80, 112, 192, 320] , _a : List[int] = [] , _a : List[int] = [1, 2, 2, 2, 1, 2, 1] , _a : List[int] = [1, 2, 2, 3, 3, 4, 1] , _a : List[int] = [1, 6, 6, 6, 6, 6, 6] , _a : float = 0.25 , _a : str = "swish" , _a : int = 2_560 , _a : str = "mean" , _a : float = 0.02 , _a : float = 0.001 , _a : float = 0.99 , _a : float = 0.2 , **_a : List[Any] , ): super().__init__(**_a ) UpperCamelCase__ = num_channels UpperCamelCase__ = image_size UpperCamelCase__ = width_coefficient UpperCamelCase__ = depth_coefficient UpperCamelCase__ = depth_divisor UpperCamelCase__ = kernel_sizes UpperCamelCase__ = in_channels UpperCamelCase__ = out_channels UpperCamelCase__ = depthwise_padding UpperCamelCase__ = strides UpperCamelCase__ = num_block_repeats UpperCamelCase__ = expand_ratios UpperCamelCase__ = squeeze_expansion_ratio UpperCamelCase__ = hidden_act UpperCamelCase__ = hidden_dim UpperCamelCase__ = pooling_type UpperCamelCase__ = initializer_range UpperCamelCase__ = batch_norm_eps UpperCamelCase__ = batch_norm_momentum UpperCamelCase__ = drop_connect_rate UpperCamelCase__ = sum(_a ) * 4 @classmethod def A_ ( cls : Tuple , _a : Union[str, os.PathLike] , **_a : Union[str, Any] ): cls._set_token_in_kwargs(_a ) UpperCamelCase__ , UpperCamelCase__ = cls.get_config_dict(_a , **_a ) # get the vision config dict if we are loading from AlignConfig if config_dict.get('''model_type''' ) == "align": UpperCamelCase__ = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(_a , **_a ) class __lowercase ( A ): '''simple docstring''' _A : List[Any] = '''align''' _A : Optional[int] = True def __init__( self : Optional[int] , _a : Tuple=None , _a : int=None , _a : Any=640 , _a : Optional[Any]=1.0 , _a : Tuple=0.02 , **_a : List[Any] , ): super().__init__(**_a ) if text_config is None: UpperCamelCase__ = {} logger.info('''text_config is None. Initializing the AlignTextConfig with default values.''' ) if vision_config is None: UpperCamelCase__ = {} logger.info('''vision_config is None. Initializing the AlignVisionConfig with default values.''' ) UpperCamelCase__ = AlignTextConfig(**_a ) UpperCamelCase__ = AlignVisionConfig(**_a ) UpperCamelCase__ = projection_dim UpperCamelCase__ = temperature_init_value UpperCamelCase__ = initializer_range @classmethod def A_ ( cls : Optional[int] , _a : AlignTextConfig , _a : AlignVisionConfig , **_a : Optional[Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_a ) def A_ ( self : Tuple ): UpperCamelCase__ = copy.deepcopy(self.__dict__ ) UpperCamelCase__ = self.text_config.to_dict() UpperCamelCase__ = self.vision_config.to_dict() UpperCamelCase__ = self.__class__.model_type return output
35
0
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) lowerCAmelCase = precision lowerCAmelCase = ceil(precision / 14 ) lowerCAmelCase = 42_68_80 * Decimal(1_00_05 ).sqrt() lowerCAmelCase = 1 lowerCAmelCase = 13_59_14_09 lowerCAmelCase = Decimal(SCREAMING_SNAKE_CASE ) for k in range(1 , SCREAMING_SNAKE_CASE ): lowerCAmelCase = factorial(6 * k ) // (factorial(3 * k ) * factorial(SCREAMING_SNAKE_CASE ) ** 3) linear_term += 5_45_14_01_34 exponential_term *= -26_25_37_41_26_40_76_80_00 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = 50 print(f'The first {n} digits of pi is: {pi(n)}')
46
"""simple docstring""" import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ = False, False, False @dataclass class lowercase : _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = None # Automatically constructed _SCREAMING_SNAKE_CASE = "dict" _SCREAMING_SNAKE_CASE = pa.struct({'bytes': pa.binary(), 'path': pa.string()} ) _SCREAMING_SNAKE_CASE = field(default='Audio' , init=_UpperCAmelCase , repr=_UpperCAmelCase ) def __call__( self ) -> Union[str, Any]: return self.pa_type def _snake_case ( self , lowercase ) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("""To support encoding audio data, please install 'soundfile'.""" ) from err if isinstance(lowercase , lowercase ): return {"bytes": None, "path": value} elif isinstance(lowercase , lowercase ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes lowerCAmelCase = BytesIO() sf.write(lowercase , value["""array"""] , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("""pcm""" ): # "PCM" only has raw audio bytes if value.get("""sampling_rate""" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("""To use PCM files, please specify a 'sampling_rate' in Audio object""" ) if value.get("""bytes""" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) lowerCAmelCase = np.frombuffer(value["""bytes"""] , dtype=np.intaa ).astype(np.floataa ) / 32_767 else: lowerCAmelCase = np.memmap(value["""path"""] , dtype="""h""" , mode="""r""" ).astype(np.floataa ) / 32_767 lowerCAmelCase = BytesIO(bytes() ) sf.write(lowercase , lowercase , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( f'An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def _snake_case ( self , lowercase , lowercase = None ) -> dict: if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Audio(decode=True) instead.""" ) lowerCAmelCase , lowerCAmelCase = (value["""path"""], BytesIO(value["""bytes"""] )) if value["""bytes"""] is not None else (value["""path"""], None) if path is None and file is None: raise ValueError(f'An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("""To support decoding audio files, please install 'librosa' and 'soundfile'.""" ) from err lowerCAmelCase = xsplitext(lowercase )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( """Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( """Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) if file is None: lowerCAmelCase = token_per_repo_id or {} lowerCAmelCase = path.split("""::""" )[-1] try: lowerCAmelCase = string_to_dict(lowercase , config.HUB_DATASETS_URL )["""repo_id"""] lowerCAmelCase = token_per_repo_id[repo_id] except (ValueError, KeyError): lowerCAmelCase = None with xopen(lowercase , """rb""" , use_auth_token=lowercase ) as f: lowerCAmelCase , lowerCAmelCase = sf.read(lowercase ) else: lowerCAmelCase , lowerCAmelCase = sf.read(lowercase ) lowerCAmelCase = array.T if self.mono: lowerCAmelCase = librosa.to_mono(lowercase ) if self.sampling_rate and self.sampling_rate != sampling_rate: lowerCAmelCase = librosa.resample(lowercase , orig_sr=lowercase , target_sr=self.sampling_rate ) lowerCAmelCase = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def _snake_case ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError("""Cannot flatten a decoded Audio feature.""" ) return { "bytes": Value("""binary""" ), "path": Value("""string""" ), } def _snake_case ( self , lowercase ) -> pa.StructArray: if pa.types.is_string(storage.type ): lowerCAmelCase = pa.array([None] * len(lowercase ) , type=pa.binary() ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): lowerCAmelCase = pa.array([None] * len(lowercase ) , type=pa.string() ) lowerCAmelCase = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("""array""" ): lowerCAmelCase = pa.array([Audio().encode_example(lowercase ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: lowerCAmelCase = storage.field("""bytes""" ) else: lowerCAmelCase = pa.array([None] * len(lowercase ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: lowerCAmelCase = storage.field("""path""" ) else: lowerCAmelCase = pa.array([None] * len(lowercase ) , type=pa.string() ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) return array_cast(lowercase , self.pa_type ) def _snake_case ( self , lowercase ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(lowercase ): with xopen(lowercase , """rb""" ) as f: lowerCAmelCase = f.read() return bytes_ lowerCAmelCase = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) lowerCAmelCase = pa.array( [os.path.basename(lowercase ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) lowerCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(lowercase , self.pa_type )
46
1
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ = 50 ) -> int: '''simple docstring''' snake_case : Union[str, Any] = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(f"{solution() = }")
83
1
'''simple docstring''' import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP lowerCAmelCase__ = False try: lowerCAmelCase__ = _is_package_available('''google.colab''') except ModuleNotFoundError: pass @input.register class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : str = None ,lowercase__ : list = [] ): __lowercase = 0 __lowercase = choices __lowercase = prompt if sys.platform == "win32": __lowercase = '''*''' else: __lowercase = '''➔ ''' def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[int] ,lowercase__ : str = "" ): if sys.platform != "win32": writeColor(self.choices[index] ,3_2 ,lowercase__ ) else: forceWrite(self.choices[index] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : int ): if index == self.position: forceWrite(F" {self.arrow_char} " ) self.write_choice(lowercase__ ) else: forceWrite(F" {self.choices[index]}" ) reset_cursor() def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Direction ,lowercase__ : int = 1 ): __lowercase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(lowercase__ ) move_cursor(lowercase__ ,direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP['''up'''] ) def SCREAMING_SNAKE_CASE ( self : Any ): self.move_direction(Direction.UP ) @input.mark(KEYMAP['''down'''] ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.move_direction(Direction.DOWN ) @input.mark(KEYMAP['''newline'''] ) def SCREAMING_SNAKE_CASE ( self : List[str] ): move_cursor(len(self.choices ) - self.position ,'''DOWN''' ) return self.position @input.mark(KEYMAP['''interrupt'''] ) def SCREAMING_SNAKE_CASE ( self : str ): move_cursor(len(self.choices ) - self.position ,'''DOWN''' ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(lowercase__ )] for number in range(1_0 )] ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = int(chr(self.current_selection ) ) __lowercase = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP ,-movement ) elif self.position < index: self.move_direction(Direction.DOWN ,lowercase__ ) else: return else: return def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : int = 0 ): if self.prompt: linebreak() forceWrite(self.prompt ,'''\n''' ) if in_colab: forceWrite('''Please input a choice index (starting from 0), and press enter''' ,'''\n''' ) else: forceWrite('''Please select a choice using the arrow or number keys, and selecting with enter''' ,'''\n''' ) __lowercase = default_choice for i in range(len(self.choices ) ): self.print_choice(lowercase__ ) forceWrite('''\n''' ) move_cursor(len(self.choices ) - self.position ,'''UP''' ) with cursor.hide(): while True: if in_colab: try: __lowercase = int(builtins.input() ) except ValueError: __lowercase = default_choice else: __lowercase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 ,'''UP''' ) clear_line() self.write_choice(lowercase__ ,'''\n''' ) return choice
104
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def __snake_case ( __UpperCamelCase : NDArray[floataa] ,__UpperCamelCase : NDArray[floataa] ,__UpperCamelCase : list[int] ,__UpperCamelCase : int ,): """simple docstring""" A_ , A_ = coefficient_matrix.shape A_ , A_ = constant_matrix.shape if rowsa != colsa: A_ = f'''Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}''' raise ValueError(__UpperCamelCase ) if colsa != 1: A_ = f'''Constant matrix must be nx1 but received {rowsa}x{colsa}''' raise ValueError(__UpperCamelCase ) if rowsa != rowsa: A_ = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " f'''received {rowsa}x{colsa} and {rowsa}x{colsa}''' ) raise ValueError(__UpperCamelCase ) if len(__UpperCamelCase ) != rowsa: A_ = ( "Number of initial values must be equal to number of rows in coefficient " f'''matrix but received {len(__UpperCamelCase )} and {rowsa}''' ) raise ValueError(__UpperCamelCase ) if iterations <= 0: raise ValueError("Iterations must be at least 1" ) A_ = np.concatenate( (coefficient_matrix, constant_matrix) ,axis=1 ) A_ , A_ = table.shape strictly_diagonally_dominant(__UpperCamelCase ) # Iterates the whole matrix for given number of times for _ in range(__UpperCamelCase ): A_ = [] for row in range(__UpperCamelCase ): A_ = 0 for col in range(__UpperCamelCase ): if col == row: A_ = table[row][col] elif col == cols - 1: A_ = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] A_ = (temp + val) / denom new_val.append(__UpperCamelCase ) A_ = new_val return [float(__UpperCamelCase ) for i in new_val] def __snake_case ( __UpperCamelCase : NDArray[floataa] ): """simple docstring""" A_ , A_ = table.shape A_ = True for i in range(0 ,__UpperCamelCase ): A_ = 0 for j in range(0 ,cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("Coefficient matrix is not strictly diagonally dominant" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
312
0
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class a : """simple docstring""" __UpperCAmelCase : Union[str, Any] = MBartConfig __UpperCAmelCase : Optional[int] = {} __UpperCAmelCase : Union[str, Any] = "gelu" def __init__( self : List[str] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Union[str, Any]=13 , lowerCamelCase : List[Any]=7 , lowerCamelCase : Dict=True , lowerCamelCase : List[Any]=False , lowerCamelCase : List[Any]=99 , lowerCamelCase : Optional[Any]=32 , lowerCamelCase : Optional[int]=2 , lowerCamelCase : List[str]=4 , lowerCamelCase : Optional[int]=37 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=20 , lowerCamelCase : List[Any]=2 , lowerCamelCase : Optional[Any]=1 , lowerCamelCase : int=0 , ) -> Union[str, Any]: __snake_case : str = parent __snake_case : List[Any] = batch_size __snake_case : Optional[Any] = seq_length __snake_case : List[Any] = is_training __snake_case : List[Any] = use_labels __snake_case : Any = vocab_size __snake_case : Optional[Any] = hidden_size __snake_case : Optional[int] = num_hidden_layers __snake_case : Optional[int] = num_attention_heads __snake_case : List[Any] = intermediate_size __snake_case : int = hidden_dropout_prob __snake_case : str = attention_probs_dropout_prob __snake_case : Tuple = max_position_embeddings __snake_case : Any = eos_token_id __snake_case : Dict = pad_token_id __snake_case : Union[str, Any] = bos_token_id def __snake_case ( self : Any ) -> List[str]: __snake_case : int = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __snake_case : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __snake_case : Any = tf.concat([input_ids, eos_tensor] , axis=1 ) __snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Optional[int] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __snake_case : Tuple = prepare_mbart_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, inputs_dict def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : Union[str, Any] ) -> Any: __snake_case : Optional[int] = TFMBartModel(config=lowerCamelCase ).get_decoder() __snake_case : List[str] = inputs_dict["input_ids"] __snake_case : List[Any] = input_ids[:1, :] __snake_case : Dict = inputs_dict["attention_mask"][:1, :] __snake_case : Dict = inputs_dict["head_mask"] __snake_case : int = 1 # first forward pass __snake_case : Any = model(lowerCamelCase , attention_mask=lowerCamelCase , head_mask=lowerCamelCase , use_cache=lowerCamelCase ) __snake_case , __snake_case : List[Any] = outputs.to_tuple() __snake_case : str = past_key_values[1] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , ): if attention_mask is None: __snake_case : Any = tf.cast(tf.math.not_equal(__lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __snake_case : List[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __snake_case : int = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __snake_case : Union[str, Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __snake_case : List[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : List[str] = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () __UpperCAmelCase : List[Any] = (TFMBartForConditionalGeneration,) if is_tf_available() else () __UpperCAmelCase : List[Any] = ( { "conversational": TFMBartForConditionalGeneration, "feature-extraction": TFMBartModel, "summarization": TFMBartForConditionalGeneration, "text2text-generation": TFMBartForConditionalGeneration, "translation": TFMBartForConditionalGeneration, } if is_tf_available() else {} ) __UpperCAmelCase : List[str] = True __UpperCAmelCase : int = False __UpperCAmelCase : Tuple = False def __snake_case ( self : Optional[int] , lowerCamelCase : Tuple , lowerCamelCase : str , lowerCamelCase : List[Any] , lowerCamelCase : Dict , lowerCamelCase : Dict ) -> Union[str, Any]: if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Dict = TFMBartModelTester(self ) __snake_case : Tuple = ConfigTester(self , config_class=lowerCamelCase ) def __snake_case ( self : Tuple ) -> int: self.config_tester.run_common_tests() def __snake_case ( self : Optional[int] ) -> Optional[Any]: __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCamelCase ) @require_sentencepiece @require_tokenizers @require_tf class a (unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = [ " UN Chief Says There Is No Military Solution in Syria", ] __UpperCAmelCase : int = [ "Şeful ONU declară că nu există o soluţie militară în Siria", ] __UpperCAmelCase : Dict = "facebook/mbart-large-en-ro" @cached_property def __snake_case ( self : Optional[int] ) -> Dict: return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def __snake_case ( self : str ) -> str: __snake_case : Dict = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def __snake_case ( self : Tuple , **lowerCamelCase : Dict ) -> Optional[Any]: __snake_case : Union[str, Any] = self.translate_src_text(**lowerCamelCase ) self.assertListEqual(self.expected_text , lowerCamelCase ) def __snake_case ( self : str , **lowerCamelCase : Optional[Any] ) -> Optional[int]: __snake_case : Optional[int] = self.tokenizer(self.src_text , **lowerCamelCase , return_tensors="tf" ) __snake_case : Any = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) __snake_case : Optional[int] = self.tokenizer.batch_decode(lowerCamelCase , skip_special_tokens=lowerCamelCase ) return generated_words @slow def __snake_case ( self : List[str] ) -> Optional[Any]: self._assert_generated_batch_equal_expected()
134
from ..utils import DummyObject, requires_backends class a (metaclass=_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : int = ["speech"] def __init__( self : List[Any] , *lowerCamelCase : List[Any] , **lowerCamelCase : Optional[Any] ) -> Dict: requires_backends(self , ["speech"] ) class a (metaclass=_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["speech"] def __init__( self : int , *lowerCamelCase : List[Any] , **lowerCamelCase : List[Any] ) -> Optional[int]: requires_backends(self , ["speech"] )
134
1
def UpperCAmelCase ( a_ ) -> int: """simple docstring""" __A = len(a_ ) __A = len(matrix[0] ) __A = min(a_ , a_ ) for row in range(a_ ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , a_ ): __A = matrix[col][row] / matrix[row][row] for i in range(a_ , a_ ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __A = True for i in range(row + 1 , a_ ): if matrix[i][row] != 0: __A , __A = matrix[i], matrix[row] __A = False break if reduce: rank -= 1 for i in range(a_ ): __A = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
15
"""simple docstring""" from collections.abc import Callable import numpy as np def __lowerCamelCase ( a_ : Callable , a_ : float , a_ : float , a_ : float , a_ : float ) -> np.ndarray: __SCREAMING_SNAKE_CASE :List[Any] = int(np.ceil((x_end - xa) / step_size ) ) __SCREAMING_SNAKE_CASE :Optional[Any] = np.zeros((n + 1,) ) __SCREAMING_SNAKE_CASE :int = ya __SCREAMING_SNAKE_CASE :str = xa for k in range(a_ ): __SCREAMING_SNAKE_CASE :Optional[int] = y[k] + step_size * ode_func(a_ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
191
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> Optional[int]: # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. snake_case_ : Optional[Any] = [[1, 2, 4], [1, 2, 3, 4]] snake_case_ : int = DisjunctiveConstraint(_SCREAMING_SNAKE_CASE ) self.assertTrue(isinstance(dc.token_ids , _SCREAMING_SNAKE_CASE ) ) with self.assertRaises(_SCREAMING_SNAKE_CASE ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(_SCREAMING_SNAKE_CASE ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def _lowerCAmelCase ( self ) -> str: # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). snake_case_ : Tuple = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(_SCREAMING_SNAKE_CASE ): DisjunctiveConstraint(_SCREAMING_SNAKE_CASE ) # fails here def _lowerCAmelCase ( self ) -> Optional[Any]: snake_case_ : Union[str, Any] = [[1, 2, 3], [1, 2, 4]] snake_case_ : Optional[int] = DisjunctiveConstraint(_SCREAMING_SNAKE_CASE ) snake_case_ : List[Any] = dc.update(1 ) snake_case_ : Optional[Any] = stepped is True and completed is False and reset is False self.assertTrue(_SCREAMING_SNAKE_CASE ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_ : Any = dc.update(2 ) snake_case_ : List[str] = stepped is True and completed is False and reset is False self.assertTrue(_SCREAMING_SNAKE_CASE ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ : Tuple = dc.update(3 ) snake_case_ : Optional[Any] = stepped is True and completed is True and reset is False self.assertTrue(_SCREAMING_SNAKE_CASE ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def _lowerCAmelCase ( self ) -> Optional[int]: snake_case_ : List[Any] = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] snake_case_ : List[Any] = DisjunctiveConstraint(_SCREAMING_SNAKE_CASE ) snake_case_ : str = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_ : str = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ : Optional[int] = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) snake_case_ : List[str] = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() snake_case_ : Optional[Any] = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) snake_case_ : str = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ : int = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
355
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A : List[Any] = ['pixel_values'] def __init__( self , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = PIL.Image.BICUBIC , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = 1 / 255 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> None: super().__init__(**_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = size if size is not None else {"height": 256, "width": 256} snake_case_ : int = get_size_dict(_SCREAMING_SNAKE_CASE ) snake_case_ : Tuple = crop_size if crop_size is not None else {"height": 224, "width": 224} snake_case_ : Dict = get_size_dict(_SCREAMING_SNAKE_CASE , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : Tuple = size snake_case_ : Tuple = resample snake_case_ : Dict = do_center_crop snake_case_ : Any = crop_size snake_case_ : int = do_rescale snake_case_ : Union[str, Any] = rescale_factor snake_case_ : Optional[int] = do_normalize snake_case_ : Union[str, Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = PIL.Image.BICUBIC , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> np.ndarray: snake_case_ : List[Any] = get_size_dict(_SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return resize( _SCREAMING_SNAKE_CASE , size=(size["height"], size["width"]) , resample=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> np.ndarray: snake_case_ : str = get_size_dict(_SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return center_crop(_SCREAMING_SNAKE_CASE , size=(size["height"], size["width"]) , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> Optional[int]: return rescale(_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> np.ndarray: return normalize(_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = ChannelDimension.FIRST , **_SCREAMING_SNAKE_CASE , ) -> PIL.Image.Image: snake_case_ : Optional[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Tuple = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Union[str, Any] = image_mean if image_mean is not None else self.image_mean snake_case_ : Optional[int] = image_std if image_std is not None else self.image_std snake_case_ : Optional[Any] = size if size is not None else self.size snake_case_ : int = get_size_dict(_SCREAMING_SNAKE_CASE ) snake_case_ : str = crop_size if crop_size is not None else self.crop_size snake_case_ : Optional[Any] = get_size_dict(_SCREAMING_SNAKE_CASE , param_name="crop_size" ) snake_case_ : int = make_list_of_images(_SCREAMING_SNAKE_CASE ) if not valid_images(_SCREAMING_SNAKE_CASE ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Optional[int] = [to_numpy_array(_SCREAMING_SNAKE_CASE ) for image in images] if do_resize: snake_case_ : Optional[Any] = [self.resize(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE ) for image in images] if do_center_crop: snake_case_ : List[Any] = [self.center_crop(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE ) for image in images] if do_rescale: snake_case_ : Optional[int] = [self.rescale(image=_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE ) for image in images] if do_normalize: snake_case_ : List[str] = [self.normalize(image=_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE ) for image in images] snake_case_ : int = [to_channel_dimension_format(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for image in images] snake_case_ : List[str] = {"pixel_values": images} return BatchFeature(data=_SCREAMING_SNAKE_CASE , tensor_type=_SCREAMING_SNAKE_CASE )
36
0
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class snake_case__(_UpperCamelCase , unittest.TestCase ): """simple docstring""" lowercase_ = LxmertTokenizer lowercase_ = LxmertTokenizerFast lowercase_ = True lowercase_ = True def snake_case ( self : List[str] ): super().setUp() lowercase__ : List[Any] = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowercase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def snake_case ( self : int , SCREAMING_SNAKE_CASE : Tuple ): lowercase__ : List[Any] = "UNwant\u00E9d,running" lowercase__ : Tuple = "unwanted, running" return input_text, output_text def snake_case ( self : List[Any] ): lowercase__ : str = self.tokenizer_class(self.vocab_file ) lowercase__ : int = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE ) , [7, 4, 5, 10, 8, 9] ) def snake_case ( self : Union[str, Any] ): if not self.test_rust_tokenizer: return lowercase__ : List[Any] = self.get_tokenizer() lowercase__ : List[Any] = self.get_rust_tokenizer() lowercase__ : Tuple = "I was born in 92000, and this is falsé." lowercase__ : Optional[Any] = tokenizer.tokenize(SCREAMING_SNAKE_CASE ) lowercase__ : Dict = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE ) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE ) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : int = self.get_rust_tokenizer() lowercase__ : List[Any] = tokenizer.encode(SCREAMING_SNAKE_CASE ) lowercase__ : Union[str, Any] = rust_tokenizer.encode(SCREAMING_SNAKE_CASE ) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )
130
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case__(_UpperCamelCase ): """simple docstring""" def snake_case ( self : Optional[Any] ): lowercase__ : str = tempfile.mkdtemp() lowercase__ : Optional[Any] = 8 # DPR tok lowercase__ : Dict = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowercase__ : List[Any] = os.path.join(self.tmpdirname , "dpr_tokenizer" ) os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) lowercase__ : str = os.path.join(SCREAMING_SNAKE_CASE , DPR_VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) # BART tok lowercase__ : Optional[Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowercase__ : List[str] = dict(zip(SCREAMING_SNAKE_CASE , range(len(SCREAMING_SNAKE_CASE ) ) ) ) lowercase__ : Dict = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowercase__ : List[Any] = {"unk_token": "<unk>"} lowercase__ : Any = os.path.join(self.tmpdirname , "bart_tokenizer" ) os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) lowercase__ : Tuple = os.path.join(SCREAMING_SNAKE_CASE , BART_VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ : str = os.path.join(SCREAMING_SNAKE_CASE , BART_VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(SCREAMING_SNAKE_CASE ) ) def snake_case ( self : Any ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def snake_case ( self : Any ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def snake_case ( self : Any ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , "bart_tokenizer" ) ) def snake_case ( self : Tuple ): shutil.rmtree(self.tmpdirname ) def snake_case ( self : Optional[int] ): lowercase__ : int = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def snake_case ( self : List[str] ): lowercase__ : Union[str, Any] = self.get_dummy_dataset() lowercase__ : str = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: lowercase__ : Union[str, Any] = dataset lowercase__ : List[str] = RagRetriever( SCREAMING_SNAKE_CASE , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE : bool ): lowercase__ : Union[str, Any] = self.get_dummy_dataset() lowercase__ : Optional[int] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="custom" , ) if from_disk: lowercase__ : Any = os.path.join(self.tmpdirname , "dataset" ) lowercase__ : Union[str, Any] = os.path.join(self.tmpdirname , "index.faiss" ) dataset.get_index("embeddings" ).save(os.path.join(self.tmpdirname , "index.faiss" ) ) dataset.drop_index("embeddings" ) dataset.save_to_disk(os.path.join(self.tmpdirname , "dataset" ) ) del dataset lowercase__ : Tuple = RagRetriever( SCREAMING_SNAKE_CASE , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowercase__ : Dict = RagRetriever( SCREAMING_SNAKE_CASE , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , SCREAMING_SNAKE_CASE ) , ) return retriever def snake_case ( self : Tuple ): lowercase__ : Optional[int] = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) lowercase__ : Union[str, Any] = os.path.join(self.tmpdirname , "hf_bert_base.hnswSQ8_correct_phi_128.c_index" ) dataset.save_faiss_index("embeddings" , index_file_name + ".index.dpr" ) pickle.dump(dataset["id"] , open(index_file_name + ".index_meta.dpr" , "wb" ) ) lowercase__ : Optional[int] = os.path.join(self.tmpdirname , "psgs_w100.tsv.pkl" ) lowercase__ : List[str] = {sample["id"]: [sample["text"], sample["title"]] for sample in dataset} pickle.dump(SCREAMING_SNAKE_CASE , open(SCREAMING_SNAKE_CASE , "wb" ) ) lowercase__ : int = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="legacy" , index_path=self.tmpdirname , ) lowercase__ : Any = RagRetriever( SCREAMING_SNAKE_CASE , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def snake_case ( self : int ): lowercase__ : Any = 1 lowercase__ : str = self.get_dummy_canonical_hf_index_retriever() lowercase__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=SCREAMING_SNAKE_CASE ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def snake_case ( self : str ): lowercase__ : Dict = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: lowercase__ : Tuple = self.get_dummy_dataset() retriever.save_pretrained(SCREAMING_SNAKE_CASE ) lowercase__ : Optional[Any] = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : List[str] = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=1 ) self.assertTrue(out is not None ) def snake_case ( self : str ): lowercase__ : Union[str, Any] = 1 lowercase__ : Tuple = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE ) lowercase__ : List[str] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ , lowercase__ , lowercase__ : Optional[Any] = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=SCREAMING_SNAKE_CASE ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def snake_case ( self : Union[str, Any] ): lowercase__ : str = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE ) lowercase__ : List[Any] = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : str = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=1 ) self.assertTrue(out is not None ) def snake_case ( self : Union[str, Any] ): lowercase__ : Optional[Any] = 1 lowercase__ : str = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ , lowercase__ , lowercase__ : Dict = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=SCREAMING_SNAKE_CASE ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def snake_case ( self : List[str] ): lowercase__ : List[str] = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE ) lowercase__ : int = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : Dict = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=1 ) self.assertTrue(out is not None ) def snake_case ( self : Union[str, Any] ): lowercase__ : List[Any] = 1 lowercase__ : List[str] = self.get_dummy_legacy_index_retriever() lowercase__ : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ , lowercase__ , lowercase__ : str = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=SCREAMING_SNAKE_CASE ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["text", "title"] ) self.assertEqual(len(doc_dicts[0]["text"] ) , SCREAMING_SNAKE_CASE ) self.assertEqual(doc_dicts[0]["text"][0] , "bar" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["text"][0] , "foo" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def snake_case ( self : Dict ): lowercase__ : Optional[int] = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : str = retriever.retrieve(SCREAMING_SNAKE_CASE , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def snake_case ( self : Any ): import torch lowercase__ : List[Any] = 1 lowercase__ : Union[str, Any] = self.get_dummy_canonical_hf_index_retriever() lowercase__ : Tuple = [[5, 7], [10, 11]] lowercase__ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : int = retriever(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE ) lowercase__ , lowercase__ , lowercase__ : List[str] = ( out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , np.ndarray ) lowercase__ : List[str] = retriever( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE , return_tensors="pt" , ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : List[Any] = ( # noqa: F841 out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], out["doc_ids"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(SCREAMING_SNAKE_CASE , torch.Tensor ) self.assertIsInstance(SCREAMING_SNAKE_CASE , torch.Tensor ) self.assertIsInstance(SCREAMING_SNAKE_CASE , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def snake_case ( self : int ): lowercase__ : List[Any] = self.get_dpr_ctx_encoder_tokenizer() lowercase__ : Optional[int] = 1 lowercase__ : str = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE ) retriever.set_ctx_encoder_tokenizer(SCREAMING_SNAKE_CASE ) lowercase__ : Union[str, Any] = [[5, 7], [10, 11]] lowercase__ : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowercase__ : List[Any] = retriever(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE ) self.assertEqual( len(SCREAMING_SNAKE_CASE ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ("tokenized_doc_ids", "tokenized_doc_attention_mask") ) , SCREAMING_SNAKE_CASE ) # check for doc token related keys in dictionary.
130
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Optional[Any] = logging.get_logger(__name__) lowercase__ : List[Any] = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : Union[str, Any] = 'biogpt' def __init__( self : Optional[Any] , lowerCAmelCase__ : List[str]=42384 , lowerCAmelCase__ : Optional[int]=1024 , lowerCAmelCase__ : List[str]=24 , lowerCAmelCase__ : List[Any]=16 , lowerCAmelCase__ : Optional[int]=4096 , lowerCAmelCase__ : Optional[int]="gelu" , lowerCAmelCase__ : Optional[Any]=0.1 , lowerCAmelCase__ : Optional[int]=0.1 , lowerCAmelCase__ : Union[str, Any]=1024 , lowerCAmelCase__ : List[str]=0.02 , lowerCAmelCase__ : Tuple=1e-1_2 , lowerCAmelCase__ : Dict=True , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : Dict=0.0 , lowerCAmelCase__ : Union[str, Any]=0.0 , lowerCAmelCase__ : Optional[int]=1 , lowerCAmelCase__ : Union[str, Any]=0 , lowerCAmelCase__ : Optional[Any]=2 , **lowerCAmelCase__ : Optional[Any] , ) -> Tuple: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = scale_embedding _UpperCamelCase = use_cache _UpperCamelCase = layerdrop _UpperCamelCase = activation_dropout super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ )
287
'''simple docstring''' import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Tuple ) -> int: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def snake_case__ ( self : int ) -> Tuple: '''simple docstring''' _UpperCamelCase = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def snake_case__ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def snake_case__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' _UpperCamelCase = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def snake_case__ ( self : Dict ) -> Dict: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', # Removed: 'text_encoder/model.safetensors', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def snake_case__ ( self : Any ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : Optional[Any] ) -> Dict: '''simple docstring''' _UpperCamelCase = [ '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] _UpperCamelCase = '''fp16''' self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : Optional[int] ) -> Dict: '''simple docstring''' _UpperCamelCase = [ '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : Optional[Any] ) -> str: '''simple docstring''' _UpperCamelCase = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def snake_case__ ( self : List[str] ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', # 'text_encoder/model.fp16.safetensors', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _UpperCamelCase = '''fp16''' self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
287
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class a__ : def __init__( self : List[str] , a : Any , a : Union[str, Any]=12 , a : List[Any]=7 , a : str=True , a : str=True , a : Dict=True , a : Union[str, Any]=99 , a : Optional[Any]=32 , a : int=32 , a : int=2 , a : Optional[int]=4 , a : Dict=37 , a : Optional[Any]=0.1 , a : Dict=0.1 , a : Optional[int]=5_12 , a : List[Any]=0.02 , a : Union[str, Any]=0 , a : Any=None , ): """simple docstring""" __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = projection_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = dropout __lowerCamelCase = attention_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = initializer_range __lowerCamelCase = scope __lowerCamelCase = bos_token_id def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: __lowerCamelCase = input_mask.numpy() __lowerCamelCase , __lowerCamelCase = input_mask.shape __lowerCamelCase = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): __lowerCamelCase = 1 __lowerCamelCase = 0 __lowerCamelCase = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : int , a : Tuple , a : Union[str, Any] ): """simple docstring""" __lowerCamelCase = TFBlipTextModel(config=a ) __lowerCamelCase = model(a , attention_mask=a , training=a ) __lowerCamelCase = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class a__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCamelCase : List[Any] =(TFBlipTextModel,) if is_tf_available() else () lowerCamelCase : Any =False lowerCamelCase : List[str] =False lowerCamelCase : int =False def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" __lowerCamelCase = BlipTextModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=a , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" pass @unittest.skip(reason='''Blip does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" pass @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def SCREAMING_SNAKE_CASE__ ( self : int , a : int=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=a )
67
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowerCamelCase : Tuple =logging.get_logger(__name__) lowerCamelCase : Optional[Any] ={ '''microsoft/conditional-detr-resnet-50''': ( '''https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json''' ), } class __a ( A__ ): _lowerCAmelCase : Optional[Any] = '''conditional_detr''' _lowerCAmelCase : List[Any] = ['''past_key_values'''] _lowerCAmelCase : List[str] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self : str , SCREAMING_SNAKE_CASE : Tuple=True , SCREAMING_SNAKE_CASE : Optional[int]=None , SCREAMING_SNAKE_CASE : Tuple=3 , SCREAMING_SNAKE_CASE : Any=3_00 , SCREAMING_SNAKE_CASE : Tuple=6 , SCREAMING_SNAKE_CASE : int=20_48 , SCREAMING_SNAKE_CASE : Union[str, Any]=8 , SCREAMING_SNAKE_CASE : int=6 , SCREAMING_SNAKE_CASE : Dict=20_48 , SCREAMING_SNAKE_CASE : Optional[int]=8 , SCREAMING_SNAKE_CASE : Dict=0.0 , SCREAMING_SNAKE_CASE : List[Any]=0.0 , SCREAMING_SNAKE_CASE : int=True , SCREAMING_SNAKE_CASE : int="relu" , SCREAMING_SNAKE_CASE : Any=2_56 , SCREAMING_SNAKE_CASE : int=0.1 , SCREAMING_SNAKE_CASE : int=0.0 , SCREAMING_SNAKE_CASE : Optional[int]=0.0 , SCREAMING_SNAKE_CASE : List[str]=0.0_2 , SCREAMING_SNAKE_CASE : Optional[int]=1.0 , SCREAMING_SNAKE_CASE : List[str]=False , SCREAMING_SNAKE_CASE : int="sine" , SCREAMING_SNAKE_CASE : str="resnet50" , SCREAMING_SNAKE_CASE : int=True , SCREAMING_SNAKE_CASE : Tuple=False , SCREAMING_SNAKE_CASE : Tuple=2 , SCREAMING_SNAKE_CASE : Tuple=5 , SCREAMING_SNAKE_CASE : List[str]=2 , SCREAMING_SNAKE_CASE : List[str]=1 , SCREAMING_SNAKE_CASE : Optional[Any]=1 , SCREAMING_SNAKE_CASE : str=2 , SCREAMING_SNAKE_CASE : Any=5 , SCREAMING_SNAKE_CASE : int=2 , SCREAMING_SNAKE_CASE : List[Any]=0.2_5 , **SCREAMING_SNAKE_CASE : Union[str, Any] , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCamelCase__ : Tuple = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase__ : Union[str, Any] = backbone_config.get("model_type" ) UpperCamelCase__ : Tuple = CONFIG_MAPPING[backbone_model_type] UpperCamelCase__ : Any = config_class.from_dict(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Optional[Any] = use_timm_backbone UpperCamelCase__ : List[Any] = backbone_config UpperCamelCase__ : Tuple = num_channels UpperCamelCase__ : List[Any] = num_queries UpperCamelCase__ : Dict = d_model UpperCamelCase__ : Any = encoder_ffn_dim UpperCamelCase__ : List[str] = encoder_layers UpperCamelCase__ : List[str] = encoder_attention_heads UpperCamelCase__ : Optional[int] = decoder_ffn_dim UpperCamelCase__ : str = decoder_layers UpperCamelCase__ : Optional[Any] = decoder_attention_heads UpperCamelCase__ : int = dropout UpperCamelCase__ : Optional[int] = attention_dropout UpperCamelCase__ : Any = activation_dropout UpperCamelCase__ : int = activation_function UpperCamelCase__ : int = init_std UpperCamelCase__ : List[Any] = init_xavier_std UpperCamelCase__ : List[str] = encoder_layerdrop UpperCamelCase__ : List[Any] = decoder_layerdrop UpperCamelCase__ : List[Any] = encoder_layers UpperCamelCase__ : Optional[Any] = auxiliary_loss UpperCamelCase__ : List[str] = position_embedding_type UpperCamelCase__ : Optional[Any] = backbone UpperCamelCase__ : Optional[int] = use_pretrained_backbone UpperCamelCase__ : List[Any] = dilation # Hungarian matcher UpperCamelCase__ : List[str] = class_cost UpperCamelCase__ : Union[str, Any] = bbox_cost UpperCamelCase__ : int = giou_cost # Loss coefficients UpperCamelCase__ : str = mask_loss_coefficient UpperCamelCase__ : List[Any] = dice_loss_coefficient UpperCamelCase__ : int = cls_loss_coefficient UpperCamelCase__ : Tuple = bbox_loss_coefficient UpperCamelCase__ : Any = giou_loss_coefficient UpperCamelCase__ : List[Any] = focal_alpha super().__init__(is_encoder_decoder=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.encoder_attention_heads @property def __lowercase ( self : Tuple ): '''simple docstring''' return self.d_model def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCamelCase__ : Any = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCamelCase__ : List[str] = self.backbone_config.to_dict() UpperCamelCase__ : int = self.__class__.model_type return output class __a ( A__ ): _lowerCAmelCase : Tuple = version.parse('''1.11''' ) @property def __lowercase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def __lowercase ( self : Tuple ): '''simple docstring''' return 1e-5 @property def __lowercase ( self : List[str] ): '''simple docstring''' return 12
189
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Optional[int] =logging.get_logger(__name__) def __lowercase ( a__ , a__=False ) -> Tuple: __SCREAMING_SNAKE_CASE = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'deit.embeddings.cls_token'), ('dist_token', 'deit.embeddings.distillation_token'), ('patch_embed.proj.weight', 'deit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'deit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'deit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" __SCREAMING_SNAKE_CASE = [(pair[0], pair[1][4:]) if pair[1].startswith('deit' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('norm.weight', 'deit.layernorm.weight'), ('norm.bias', 'deit.layernorm.bias'), ('head.weight', 'cls_classifier.weight'), ('head.bias', 'cls_classifier.bias'), ('head_dist.weight', 'distillation_classifier.weight'), ('head_dist.bias', 'distillation_classifier.bias'), ] ) return rename_keys def __lowercase ( a__ , a__ , a__=False ) -> Tuple: for i in range(config.num_hidden_layers ): if base_model: __SCREAMING_SNAKE_CASE = '' else: __SCREAMING_SNAKE_CASE = 'deit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __SCREAMING_SNAKE_CASE = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) __SCREAMING_SNAKE_CASE = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __SCREAMING_SNAKE_CASE = in_proj_weight[ : config.hidden_size, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[: config.hidden_size] __SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __SCREAMING_SNAKE_CASE = in_proj_weight[ -config.hidden_size :, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[-config.hidden_size :] def __lowercase ( a__ , a__ , a__ ) -> str: __SCREAMING_SNAKE_CASE = dct.pop(a__ ) __SCREAMING_SNAKE_CASE = val def __lowercase ( ) -> List[Any]: __SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' __SCREAMING_SNAKE_CASE = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def __lowercase ( a__ , a__ ) -> Dict: __SCREAMING_SNAKE_CASE = DeiTConfig() # all deit models have fine-tuned heads __SCREAMING_SNAKE_CASE = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size __SCREAMING_SNAKE_CASE = 10_00 __SCREAMING_SNAKE_CASE = 'huggingface/label-files' __SCREAMING_SNAKE_CASE = 'imagenet-1k-id2label.json' __SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(a__ , a__ , repo_type='dataset' ) , 'r' ) ) __SCREAMING_SNAKE_CASE = {int(a__ ): v for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = idalabel __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = int(deit_name[-6:-4] ) __SCREAMING_SNAKE_CASE = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('tiny' ): __SCREAMING_SNAKE_CASE = 1_92 __SCREAMING_SNAKE_CASE = 7_68 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 3 elif deit_name[9:].startswith('small' ): __SCREAMING_SNAKE_CASE = 3_84 __SCREAMING_SNAKE_CASE = 15_36 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 6 if deit_name[9:].startswith('base' ): pass elif deit_name[4:].startswith('large' ): __SCREAMING_SNAKE_CASE = 10_24 __SCREAMING_SNAKE_CASE = 40_96 __SCREAMING_SNAKE_CASE = 24 __SCREAMING_SNAKE_CASE = 16 # load original model from timm __SCREAMING_SNAKE_CASE = timm.create_model(a__ , pretrained=a__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys __SCREAMING_SNAKE_CASE = timm_model.state_dict() __SCREAMING_SNAKE_CASE = create_rename_keys(a__ , a__ ) for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) read_in_q_k_v(a__ , a__ , a__ ) # load HuggingFace model __SCREAMING_SNAKE_CASE = DeiTForImageClassificationWithTeacher(a__ ).eval() model.load_state_dict(a__ ) # Check outputs on an image, prepared by DeiTImageProcessor __SCREAMING_SNAKE_CASE = int( (2_56 / 2_24) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 __SCREAMING_SNAKE_CASE = DeiTImageProcessor(size=a__ , crop_size=config.image_size ) __SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() , return_tensors='pt' ) __SCREAMING_SNAKE_CASE = encoding['pixel_values'] __SCREAMING_SNAKE_CASE = model(a__ ) __SCREAMING_SNAKE_CASE = timm_model(a__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(a__ , outputs.logits , atol=1E-3 ) Path(a__ ).mkdir(exist_ok=a__ ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a__ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a__ ) if __name__ == "__main__": lowerCAmelCase__ : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase__ : str =parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
118
from collections import deque from .hash_table import HashTable class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' def __init__( self , *_A , **_A ): '''simple docstring''' super().__init__(*_A , **_A ) def _A ( self , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_A ) __SCREAMING_SNAKE_CASE = self.values[key] def _A ( self ): '''simple docstring''' return ( sum(self.charge_factor - len(_A ) for slot in self.values ) / self.size_table * self.charge_factor ) def _A ( self , _A , _A=None ): '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_A ) == 0 ): return key return super()._collision_resolution(_A , _A )
118
1
import os import string import sys _UpperCamelCase = 1 << 8 _UpperCamelCase = { "tab": ord("\t"), "newline": ord("\r"), "esc": 27, "up": 65 + ARROW_KEY_FLAG, "down": 66 + ARROW_KEY_FLAG, "right": 67 + ARROW_KEY_FLAG, "left": 68 + ARROW_KEY_FLAG, "mod_int": 91, "undefined": sys.maxsize, "interrupt": 3, "insert": 50, "delete": 51, "pg_up": 53, "pg_down": 54, } _UpperCamelCase = KEYMAP["up"] _UpperCamelCase = KEYMAP["left"] if sys.platform == "win32": _UpperCamelCase = [] _UpperCamelCase = { B"\xe0H": KEYMAP["up"] - ARROW_KEY_FLAG, B"\x00H": KEYMAP["up"] - ARROW_KEY_FLAG, B"\xe0P": KEYMAP["down"] - ARROW_KEY_FLAG, B"\x00P": KEYMAP["down"] - ARROW_KEY_FLAG, B"\xe0M": KEYMAP["right"] - ARROW_KEY_FLAG, B"\x00M": KEYMAP["right"] - ARROW_KEY_FLAG, B"\xe0K": KEYMAP["left"] - ARROW_KEY_FLAG, B"\x00K": KEYMAP["left"] - ARROW_KEY_FLAG, } for i in range(10): _UpperCamelCase = ord(str(i)) def _lowercase ( ): if os.name == "nt": import msvcrt __lowerCAmelCase : Tuple = '''mbcs''' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(lowercase__ ) == 0: # Read the keystroke __lowerCAmelCase : str = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): __lowerCAmelCase : List[str] = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: __lowerCAmelCase : Union[str, Any] = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['''mod_int'''] ) ) WIN_CH_BUFFER.append(lowercase__ ) if ord(lowercase__ ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(1_2_6 ) ) __lowerCAmelCase : str = chr(KEYMAP['''esc'''] ) except KeyError: __lowerCAmelCase : List[str] = cha[1] else: __lowerCAmelCase : Union[str, Any] = ch.decode(lowercase__ ) else: __lowerCAmelCase : Optional[Any] = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty __lowerCAmelCase : Optional[int] = sys.stdin.fileno() __lowerCAmelCase : Tuple = termios.tcgetattr(lowercase__ ) try: tty.setraw(lowercase__ ) __lowerCAmelCase : int = sys.stdin.read(1 ) finally: termios.tcsetattr(lowercase__ , termios.TCSADRAIN , lowercase__ ) return ch def _lowercase ( ): __lowerCAmelCase : Tuple = get_raw_chars() if ord(lowercase__ ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(lowercase__ ) == KEYMAP["esc"]: __lowerCAmelCase : Tuple = get_raw_chars() if ord(lowercase__ ) == KEYMAP["mod_int"]: __lowerCAmelCase : int = get_raw_chars() if ord(lowercase__ ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(lowercase__ ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(lowercase__ ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
275
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = self.dummy_uncond_unet __lowerCAmelCase : Any = PNDMScheduler() __lowerCAmelCase : Dict = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' ).images __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''google/ddpm-cifar10-32''' __lowerCAmelCase : Union[str, Any] = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : int = PNDMScheduler() __lowerCAmelCase : Any = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
1
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def lowerCamelCase__ ( a__ : int ) -> List[Any]: UpperCamelCase_ = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """decoder.output_projection.weight""", """_float_tensor""", """encoder.embed_positions._float_tensor""", """decoder.embed_positions._float_tensor""", ] for k in ignore_keys: state_dict.pop(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase__ ( a__ : Any ) -> int: UpperCamelCase_ = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: UpperCamelCase_ = s_dict.pop(UpperCamelCase__ ) elif "subsample" in key: UpperCamelCase_ = s_dict.pop(UpperCamelCase__ ) def lowerCamelCase__ ( a__ : Tuple ) -> Union[str, Any]: UpperCamelCase_ , UpperCamelCase_ = emb.weight.shape UpperCamelCase_ = nn.Linear(UpperCamelCase__ , UpperCamelCase__ , bias=UpperCamelCase__ ) UpperCamelCase_ = emb.weight.data return lin_layer def lowerCamelCase__ ( a__ : Dict , a__ : List[str] ) -> str: UpperCamelCase_ = torch.load(UpperCamelCase__ , map_location="""cpu""" ) UpperCamelCase_ = mam_aaa["""args"""] UpperCamelCase_ = mam_aaa["""model"""] UpperCamelCase_ = state_dict["""decoder.output_projection.weight"""] remove_ignore_keys_(UpperCamelCase__ ) rename_keys(UpperCamelCase__ ) UpperCamelCase_ = state_dict["""decoder.embed_tokens.weight"""].shape[0] UpperCamelCase_ = args.share_decoder_input_output_embed UpperCamelCase_ = [int(UpperCamelCase__ ) for i in args.conv_kernel_sizes.split(""",""" )] UpperCamelCase_ = SpeechaTextConfig( vocab_size=UpperCamelCase__ , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="""relu""" , num_conv_layers=len(UpperCamelCase__ ) , conv_channels=args.conv_channels , conv_kernel_sizes=UpperCamelCase__ , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=UpperCamelCase__ , num_beams=5 , max_length=200 , use_cache=UpperCamelCase__ , decoder_start_token_id=2 , early_stopping=UpperCamelCase__ , ) UpperCamelCase_ = SpeechaTextForConditionalGeneration(UpperCamelCase__ ) UpperCamelCase_ , UpperCamelCase_ = model.model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0 and not set(UpperCamelCase__ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( """Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,""" f''' but all the following weights are missing {missing}''' ) if tie_embeds: UpperCamelCase_ = make_linear_from_emb(model.model.decoder.embed_tokens ) else: UpperCamelCase_ = lm_head_weights model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument('''--fairseq_path''', type=str, help='''Path to the fairseq model (.pt) file.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _A = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
367
import math def lowerCamelCase__ ( a__ : float , a__ : float ) -> float: if ( not isinstance(a__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("""power_factor must be a valid float value between -1 and 1.""" ) return apparent_power * power_factor def lowerCamelCase__ ( a__ : float , a__ : float ) -> float: if ( not isinstance(a__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("""power_factor must be a valid float value between -1 and 1.""" ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
261
0
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers lowerCAmelCase__ = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Tuple = os.path.dirname(os.path.realpath(A_ ) ) _lowerCamelCase : Optional[Any] = os.path.join(A_, '''words.txt''' ) _lowerCamelCase : Dict = '''''' with open(A_ ) as f: _lowerCamelCase : Any = f.readline() _lowerCamelCase : List[str] = [word.strip('''"''' ) for word in words.strip('''\r\n''' ).split(''',''' )] _lowerCamelCase : Union[str, Any] = [ word for word in [sum(ord(A_ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(A_ ) if __name__ == "__main__": print(solution())
72
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
def lowerCamelCase_ ( UpperCamelCase__ : int = 100_0000 ): '''simple docstring''' UpperCamelCase__ = set(range(3, UpperCamelCase__, 2 ) ) primes.add(2 ) for p in range(3, UpperCamelCase__, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, UpperCamelCase__, UpperCamelCase__ ) ) ) UpperCamelCase__ = [float(UpperCamelCase__ ) for n in range(limit + 1 )] for p in primes: for n in range(UpperCamelCase__, limit + 1, UpperCamelCase__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
35
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger lowercase = get_logger(__name__) class __lowercase : '''simple docstring''' def __init__( self : Dict , _a : Optional[str] = None ): UpperCamelCase__ = ( os.path.join(_a , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) UpperCamelCase__ = Extractor def A_ ( self : str , _a : str ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" UpperCamelCase__ = os.path.abspath(_a ) return os.path.join(self.extract_dir , hash_url_to_filename(_a ) ) def A_ ( self : Optional[Any] , _a : str , _a : bool ): return force_extract or ( not os.path.isfile(_a ) and not (os.path.isdir(_a ) and os.listdir(_a )) ) def A_ ( self : int , _a : str , _a : bool = False ): UpperCamelCase__ = self.extractor.infer_extractor_format(_a ) if not extractor_format: return input_path UpperCamelCase__ = self._get_output_path(_a ) if self._do_extract(_a , _a ): self.extractor.extract(_a , _a , _a ) return output_path class __lowercase ( A ): '''simple docstring''' @classmethod @abstractmethod def A_ ( cls : List[Any] , _a : Union[Path, str] , **_a : List[str] ): ... @staticmethod @abstractmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): ... class __lowercase ( A, A ): '''simple docstring''' _A : List[bytes] = [] @staticmethod def A_ ( _a : Union[Path, str] , _a : int ): with open(_a , '''rb''' ) as f: return f.read(_a ) @classmethod def A_ ( cls : str , _a : Union[Path, str] , _a : bytes = b"" ): if not magic_number: UpperCamelCase__ = max(len(_a ) for cls_magic_number in cls.magic_numbers ) try: UpperCamelCase__ = cls.read_magic_number(_a , _a ) except OSError: return False return any(magic_number.startswith(_a ) for cls_magic_number in cls.magic_numbers ) class __lowercase ( A ): '''simple docstring''' @classmethod def A_ ( cls : Union[str, Any] , _a : Union[Path, str] , **_a : Any ): return tarfile.is_tarfile(_a ) @staticmethod def A_ ( _a : int , _a : List[str] ): def resolved(_a : str ) -> str: return os.path.realpath(os.path.abspath(_a ) ) def badpath(_a : str , _a : str ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(_a , _a ) ).startswith(_a ) def badlink(_a : Tuple , _a : str ) -> bool: # Links are interpreted relative to the directory containing the link UpperCamelCase__ = resolved(os.path.join(_a , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=_a ) UpperCamelCase__ = resolved(_a ) for finfo in members: if badpath(finfo.name , _a ): logger.error(F"""Extraction of {finfo.name} is blocked (illegal path)""" ) elif finfo.issym() and badlink(_a , _a ): logger.error(F"""Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}""" ) elif finfo.islnk() and badlink(_a , _a ): logger.error(F"""Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}""" ) else: yield finfo @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): os.makedirs(_a , exist_ok=_a ) UpperCamelCase__ = tarfile.open(_a ) tar_file.extractall(_a , members=TarExtractor.safemembers(_a , _a ) ) tar_file.close() class __lowercase ( A ): '''simple docstring''' _A : int = [b'''\x1F\x8B'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): with gzip.open(_a , '''rb''' ) as gzip_file: with open(_a , '''wb''' ) as extracted_file: shutil.copyfileobj(_a , _a ) class __lowercase ( A ): '''simple docstring''' _A : int = [ b'''PK\x03\x04''', b'''PK\x05\x06''', # empty archive b'''PK\x07\x08''', # spanned archive ] @classmethod def A_ ( cls : Dict , _a : Union[Path, str] , _a : bytes = b"" ): if super().is_extractable(_a , magic_number=_a ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(_a , '''rb''' ) as fp: UpperCamelCase__ = _EndRecData(_a ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: UpperCamelCase__ = fp.read(_a ) # CD is where we expect it to be if len(_a ) == sizeCentralDir: UpperCamelCase__ = struct.unpack(_a , _a ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): os.makedirs(_a , exist_ok=_a ) with zipfile.ZipFile(_a , '''r''' ) as zip_file: zip_file.extractall(_a ) zip_file.close() class __lowercase ( A ): '''simple docstring''' _A : Tuple = [b'''\xFD\x37\x7A\x58\x5A\x00'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): with lzma.open(_a ) as compressed_file: with open(_a , '''wb''' ) as extracted_file: shutil.copyfileobj(_a , _a ) class __lowercase ( A ): '''simple docstring''' _A : Union[str, Any] = [b'''Rar!\x1a\x07\x00''', b'''Rar!\x1a\x07\x01\x00'''] # RAR_ID # RAR5_ID @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): if not config.RARFILE_AVAILABLE: raise ImportError('''Please pip install rarfile''' ) import rarfile os.makedirs(_a , exist_ok=_a ) UpperCamelCase__ = rarfile.RarFile(_a ) rf.extractall(_a ) rf.close() class __lowercase ( A ): '''simple docstring''' _A : Optional[Any] = [b'''\x28\xb5\x2F\xFD'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): if not config.ZSTANDARD_AVAILABLE: raise ImportError('''Please pip install zstandard''' ) import zstandard as zstd UpperCamelCase__ = zstd.ZstdDecompressor() with open(_a , '''rb''' ) as ifh, open(_a , '''wb''' ) as ofh: dctx.copy_stream(_a , _a ) class __lowercase ( A ): '''simple docstring''' _A : Any = [b'''\x42\x5A\x68'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): with bza.open(_a , '''rb''' ) as compressed_file: with open(_a , '''wb''' ) as extracted_file: shutil.copyfileobj(_a , _a ) class __lowercase ( A ): '''simple docstring''' _A : Optional[int] = [b'''\x37\x7A\xBC\xAF\x27\x1C'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): if not config.PY7ZR_AVAILABLE: raise ImportError('''Please pip install py7zr''' ) import pyazr os.makedirs(_a , exist_ok=_a ) with pyazr.SevenZipFile(_a , '''r''' ) as archive: archive.extractall(_a ) class __lowercase ( A ): '''simple docstring''' _A : Union[str, Any] = [b'''\x04\x22\x4D\x18'''] @staticmethod def A_ ( _a : Union[Path, str] , _a : Union[Path, str] ): if not config.LZ4_AVAILABLE: raise ImportError('''Please pip install lz4''' ) import lza.frame with lza.frame.open(_a , '''rb''' ) as compressed_file: with open(_a , '''wb''' ) as extracted_file: shutil.copyfileobj(_a , _a ) class __lowercase : '''simple docstring''' _A : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def A_ ( cls : Dict ): return max( len(_a ) for extractor in cls.extractors.values() if issubclass(_a , _a ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def A_ ( _a : Union[Path, str] , _a : int ): try: return MagicNumberBaseExtractor.read_magic_number(_a , magic_number_length=_a ) except OSError: return b"" @classmethod def A_ ( cls : Optional[Any] , _a : Union[Path, str] , _a : bool = False ): warnings.warn( '''Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'infer_extractor_format\' instead.''' , category=_a , ) UpperCamelCase__ = cls.infer_extractor_format(_a ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def A_ ( cls : str , _a : Union[Path, str] ): # <Added version="2.4.0"/> UpperCamelCase__ = cls._get_magic_number_max_length() UpperCamelCase__ = cls._read_magic_number(_a , _a ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(_a , magic_number=_a ): return extractor_format @classmethod def A_ ( cls : List[Any] , _a : Union[Path, str] , _a : Union[Path, str] , _a : Optional[str] = None , _a : Optional[BaseExtractor] = "deprecated" , ): os.makedirs(os.path.dirname(_a ) , exist_ok=_a ) # Prevent parallel extractions UpperCamelCase__ = str(Path(_a ).with_suffix('''.lock''' ) ) with FileLock(_a ): shutil.rmtree(_a , ignore_errors=_a ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(_a , _a ): # passed as positional arg warnings.warn( '''Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'extractor_format\' instead.''' , category=_a , ) UpperCamelCase__ = extractor if extractor != '''deprecated''' else extractor_format else: UpperCamelCase__ = cls.extractors[extractor_format] return extractor.extract(_a , _a ) else: warnings.warn( '''Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an ''' '''exception in 3.0.0.''' , category=_a , ) for extractor in cls.extractors.values(): if extractor.is_extractable(_a ): return extractor.extract(_a , _a )
35
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version a__ : Dict = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.8.0''', '''To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt''') @dataclass class UpperCamelCase_ : """simple docstring""" snake_case__ : Optional[str] = field( default="cifar10" , metadata={"help": "Name of a dataset from the datasets package"}) snake_case__ : Optional[str] = field( default=UpperCamelCase , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."}) snake_case__ : Optional[str] = field( default=UpperCamelCase , metadata={"help": "The column name of the images in the files."}) snake_case__ : Optional[str] = field(default=UpperCamelCase , metadata={"help": "A folder containing the training data."}) snake_case__ : Optional[str] = field(default=UpperCamelCase , metadata={"help": "A folder containing the validation data."}) snake_case__ : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."}) snake_case__ : Optional[int] = field( default=UpperCamelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) snake_case__ : Optional[int] = field( default=UpperCamelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: __SCREAMING_SNAKE_CASE = {} if self.train_dir is not None: __SCREAMING_SNAKE_CASE = self.train_dir if self.validation_dir is not None: __SCREAMING_SNAKE_CASE = self.validation_dir __SCREAMING_SNAKE_CASE = data_files if data_files else None @dataclass class UpperCamelCase_ : """simple docstring""" snake_case__ : str = field( default=UpperCamelCase , metadata={ "help": ( "The model checkpoint for weights initialization.Don't set if you want to train a model from scratch." ) } , ) snake_case__ : Optional[str] = field( default=UpperCamelCase , metadata={"help": "Pretrained config name or path if not the same as model_name_or_path"}) snake_case__ : Optional[str] = field( default=UpperCamelCase , metadata={ "help": ( "Override some existing default config settings when a model is trained from scratch. Example: " "n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index" ) } , ) snake_case__ : Optional[str] = field( default=UpperCamelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"}) snake_case__ : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) snake_case__ : str = field(default=UpperCamelCase , metadata={"help": "Name or path of preprocessor config."}) snake_case__ : bool = field( default=UpperCamelCase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) snake_case__ : float = field( default=0.75 , metadata={"help": "The ratio of the number of masked tokens in the input sequence."}) snake_case__ : bool = field( default=UpperCamelCase , metadata={"help": "Whether or not to train with normalized pixel values as target."}) @dataclass class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : float = field( default=1E-3 , metadata={"help": "Base learning rate: absolute_lr = base_lr * total_batch_size / 256."}) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = torch.stack([example["pixel_values"] for example in examples] ) return {"pixel_values": pixel_values} def UpperCAmelCase__ (): '''simple docstring''' __SCREAMING_SNAKE_CASE = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_mae" , lowerCAmelCase_ , lowerCAmelCase_ ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() __SCREAMING_SNAKE_CASE = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase_ ) transformers.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. __SCREAMING_SNAKE_CASE = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __SCREAMING_SNAKE_CASE = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Initialize our dataset. __SCREAMING_SNAKE_CASE = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. __SCREAMING_SNAKE_CASE = None if "validation" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowerCAmelCase_ ) and data_args.train_val_split > 0.0: __SCREAMING_SNAKE_CASE = ds["train"].train_test_split(data_args.train_val_split ) __SCREAMING_SNAKE_CASE = split["train"] __SCREAMING_SNAKE_CASE = split["test"] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __SCREAMING_SNAKE_CASE = { "cache_dir": model_args.cache_dir, "revision": model_args.model_revision, "use_auth_token": True if model_args.use_auth_token else None, } if model_args.config_name: __SCREAMING_SNAKE_CASE = ViTMAEConfig.from_pretrained(model_args.config_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: __SCREAMING_SNAKE_CASE = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: __SCREAMING_SNAKE_CASE = ViTMAEConfig() logger.warning("You are instantiating a new config instance from scratch." ) if model_args.config_overrides is not None: logger.info(f"""Overriding config: {model_args.config_overrides}""" ) config.update_from_string(model_args.config_overrides ) logger.info(f"""New config: {config}""" ) # adapt config config.update( { "mask_ratio": model_args.mask_ratio, "norm_pix_loss": model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: __SCREAMING_SNAKE_CASE = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: __SCREAMING_SNAKE_CASE = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: __SCREAMING_SNAKE_CASE = ViTImageProcessor() # create model if model_args.model_name_or_path: __SCREAMING_SNAKE_CASE = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("Training new model from scratch" ) __SCREAMING_SNAKE_CASE = ViTMAEForPreTraining(lowerCAmelCase_ ) if training_args.do_train: __SCREAMING_SNAKE_CASE = ds["train"].column_names else: __SCREAMING_SNAKE_CASE = ds["validation"].column_names if data_args.image_column_name is not None: __SCREAMING_SNAKE_CASE = data_args.image_column_name elif "image" in column_names: __SCREAMING_SNAKE_CASE = "image" elif "img" in column_names: __SCREAMING_SNAKE_CASE = "img" else: __SCREAMING_SNAKE_CASE = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: __SCREAMING_SNAKE_CASE = image_processor.size["shortest_edge"] else: __SCREAMING_SNAKE_CASE = (image_processor.size["height"], image_processor.size["width"]) __SCREAMING_SNAKE_CASE = Compose( [ Lambda(lambda lowerCAmelCase_ : img.convert("RGB" ) if img.mode != "RGB" else img ), RandomResizedCrop(lowerCAmelCase_ , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = [transforms(lowerCAmelCase_ ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: __SCREAMING_SNAKE_CASE = ds["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(lowerCAmelCase_ ) if training_args.do_eval: if "validation" not in ds: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: __SCREAMING_SNAKE_CASE = ( ds["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(lowerCAmelCase_ ) # Compute absolute learning rate __SCREAMING_SNAKE_CASE = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: __SCREAMING_SNAKE_CASE = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer __SCREAMING_SNAKE_CASE = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=ds["train"] if training_args.do_train else None , eval_dataset=ds["validation"] if training_args.do_eval else None , tokenizer=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , ) # Training if training_args.do_train: __SCREAMING_SNAKE_CASE = None if training_args.resume_from_checkpoint is not None: __SCREAMING_SNAKE_CASE = training_args.resume_from_checkpoint elif last_checkpoint is not None: __SCREAMING_SNAKE_CASE = last_checkpoint __SCREAMING_SNAKE_CASE = trainer.train(resume_from_checkpoint=lowerCAmelCase_ ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: __SCREAMING_SNAKE_CASE = trainer.evaluate() trainer.log_metrics("eval" , lowerCAmelCase_ ) trainer.save_metrics("eval" , lowerCAmelCase_ ) # Write model card and (optionally) push to hub __SCREAMING_SNAKE_CASE = { "tasks": "masked-auto-encoding", "dataset": data_args.dataset_name, "tags": ["masked-auto-encoding"], } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase_ ) else: trainer.create_model_card(**lowerCAmelCase_ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' main() if __name__ == "__main__": main()
54
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if number > 0: raise ValueError("input must be a negative integer" ) __SCREAMING_SNAKE_CASE = len(bin(lowerCAmelCase_ )[3:] ) __SCREAMING_SNAKE_CASE = bin(abs(lowerCAmelCase_ ) - (1 << binary_number_length) )[3:] __SCREAMING_SNAKE_CASE = ( ( "1" + "0" * (binary_number_length - len(lowerCAmelCase_ )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
54
1
from __future__ import annotations import time lowercase_ = list[tuple[int, int]] lowercase_ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] lowercase_ = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class A : """simple docstring""" def __init__( self : List[str],lowercase_ : int,lowercase_ : int,lowercase_ : int,lowercase_ : int,lowercase_ : Node | None )-> List[Any]: '''simple docstring''' A__ = pos_x A__ = pos_y A__ = (pos_y, pos_x) A__ = goal_x A__ = goal_y A__ = parent class A : """simple docstring""" def __init__( self : Any,lowercase_ : tuple[int, int],lowercase_ : tuple[int, int] )-> Tuple: '''simple docstring''' A__ = Node(start[1],start[0],goal[1],goal[0],lowercase_ ) A__ = Node(goal[1],goal[0],goal[1],goal[0],lowercase_ ) A__ = [self.start] A__ = False def snake_case__ ( self : int )-> Path | None: '''simple docstring''' while self.node_queue: A__ = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: A__ = True return self.retrace_path(lowercase_ ) A__ = self.get_successors(lowercase_ ) for node in successors: self.node_queue.append(lowercase_ ) if not self.reached: return [self.start.pos] return None def snake_case__ ( self : int,lowercase_ : Node )-> list[Node]: '''simple docstring''' A__ = [] for action in delta: A__ = parent.pos_x + action[1] A__ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(lowercase_,lowercase_,self.target.pos_y,self.target.pos_x,lowercase_ ) ) return successors def snake_case__ ( self : Dict,lowercase_ : Node | None )-> Path: '''simple docstring''' A__ = node A__ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) A__ = current_node.parent path.reverse() return path class A : """simple docstring""" def __init__( self : Tuple,lowercase_ : Tuple,lowercase_ : Tuple )-> str: '''simple docstring''' A__ = BreadthFirstSearch(lowercase_,lowercase_ ) A__ = BreadthFirstSearch(lowercase_,lowercase_ ) A__ = False def snake_case__ ( self : Tuple )-> Path | None: '''simple docstring''' while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: A__ = self.fwd_bfs.node_queue.pop(0 ) A__ = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: A__ = True return self.retrace_bidirectional_path( lowercase_,lowercase_ ) A__ = current_bwd_node A__ = current_fwd_node A__ = { self.fwd_bfs: self.fwd_bfs.get_successors(lowercase_ ), self.bwd_bfs: self.bwd_bfs.get_successors(lowercase_ ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(lowercase_ ) if not self.reached: return [self.fwd_bfs.start.pos] return None def snake_case__ ( self : Optional[Any],lowercase_ : Node,lowercase_ : Node )-> Path: '''simple docstring''' A__ = self.fwd_bfs.retrace_path(lowercase_ ) A__ = self.bwd_bfs.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() A__ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() lowercase_ = (0, 0) lowercase_ = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) lowercase_ = time.time() lowercase_ = BreadthFirstSearch(init, goal) lowercase_ = bfs.search() lowercase_ = time.time() - start_bfs_time print("Unidirectional BFS computation time : ", bfs_time) lowercase_ = time.time() lowercase_ = BidirectionalBreadthFirstSearch(init, goal) lowercase_ = bd_bfs.search() lowercase_ = time.time() - start_bd_bfs_time print("Bidirectional BFS computation time : ", bd_bfs_time)
282
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class A ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : List[Any],lowercase_ : str )-> List[Any]: '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'],model_result['ss'] ): A__ = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(lowercase_ ) def snake_case__ ( self : Dict )-> List[str]: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Dict )-> List[str]: '''simple docstring''' A__ = 'sgugger/tiny-distilbert-classification' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,only_pretrain_model=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : List[Any] )-> Any: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,torchscript=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == 'cpu','Cant do half precision' ) def snake_case__ ( self : Any )-> Dict: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,fpaa=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Any )-> Optional[Any]: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = AutoConfig.from_pretrained(lowercase_ ) # set architectures equal to `None` A__ = None A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_,configs=[config] ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Union[str, Any] )-> int: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == 'cpu','Can\'t do half precision' ) def snake_case__ ( self : List[Any] )-> Dict: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],fpaa=lowercase_,multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : int )-> Optional[int]: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = AutoConfig.from_pretrained(lowercase_ ) A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_,configs=[config] ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : List[Any] )-> Any: '''simple docstring''' A__ = 'sshleifer/tinier_bart' A__ = AutoConfig.from_pretrained(lowercase_ ) A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_,configs=[config] ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : List[str] )-> List[str]: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' A__ = AutoConfig.from_pretrained(lowercase_ ) A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_,configs=[config] ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : int )-> Union[str, Any]: '''simple docstring''' A__ = 'sshleifer/tinier_bart' A__ = AutoConfig.from_pretrained(lowercase_ ) A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_,configs=[config] ) A__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : Optional[Any] )-> Tuple: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,save_to_csv=lowercase_,sequence_lengths=[8],batch_sizes=[1],inference_time_csv_file=os.path.join(lowercase_,'inf_time.csv' ),train_memory_csv_file=os.path.join(lowercase_,'train_mem.csv' ),inference_memory_csv_file=os.path.join(lowercase_,'inf_mem.csv' ),train_time_csv_file=os.path.join(lowercase_,'train_time.csv' ),env_info_csv_file=os.path.join(lowercase_,'env.csv' ),multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) benchmark.run() self.assertTrue(Path(os.path.join(lowercase_,'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase_,'train_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase_,'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase_,'train_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase_,'env.csv' ) ).exists() ) def snake_case__ ( self : Tuple )-> str: '''simple docstring''' A__ = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(lowercase_ : Optional[Any] ): self.assertTrue(hasattr(lowercase_,'sequential' ) ) self.assertTrue(hasattr(lowercase_,'cumulative' ) ) self.assertTrue(hasattr(lowercase_,'current' ) ) self.assertTrue(hasattr(lowercase_,'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: A__ = PyTorchBenchmarkArguments( models=[MODEL_ID],training=lowercase_,inference=lowercase_,sequence_lengths=[8],batch_sizes=[1],log_filename=os.path.join(lowercase_,'log.txt' ),log_print=lowercase_,trace_memory_line_by_line=lowercase_,multi_process=lowercase_,) A__ = PyTorchBenchmark(lowercase_ ) A__ = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(lowercase_,'log.txt' ) ).exists() )
282
1
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _snake_case : '''simple docstring''' def __init__( self: Dict ,lowerCamelCase_: List[Any] ,lowerCamelCase_: List[str]=13 ,lowerCamelCase_: int=30 ,lowerCamelCase_: Optional[int]=2 ,lowerCamelCase_: Union[str, Any]=3 ,lowerCamelCase_: Optional[Any]=True ,lowerCamelCase_: Any=True ,lowerCamelCase_: str=32 ,lowerCamelCase_: int=5 ,lowerCamelCase_: Optional[int]=4 ,lowerCamelCase_: Any=37 ,lowerCamelCase_: List[str]="gelu" ,lowerCamelCase_: Tuple=0.1 ,lowerCamelCase_: Dict=0.1 ,lowerCamelCase_: Any=10 ,lowerCamelCase_: Dict=0.0_2 ,lowerCamelCase_: Union[str, Any]=3 ,lowerCamelCase_: List[Any]=None ,lowerCamelCase_: str=2 ,) -> Union[str, Any]: UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = batch_size UpperCAmelCase_ : List[Any] = image_size UpperCAmelCase_ : int = patch_size UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : str = is_training UpperCAmelCase_ : Any = use_labels UpperCAmelCase_ : int = hidden_size UpperCAmelCase_ : str = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : int = intermediate_size UpperCAmelCase_ : Optional[Any] = hidden_act UpperCAmelCase_ : Any = hidden_dropout_prob UpperCAmelCase_ : Tuple = attention_probs_dropout_prob UpperCAmelCase_ : List[Any] = type_sequence_label_size UpperCAmelCase_ : int = initializer_range UpperCAmelCase_ : Optional[Any] = scope UpperCAmelCase_ : Optional[int] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase_ : Optional[Any] = (image_size // patch_size) ** 2 UpperCAmelCase_ : Union[str, Any] = num_patches + 2 def A__ ( self: List[Any] ) -> Dict: UpperCAmelCase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : str = None if self.use_labels: UpperCAmelCase_ : int = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : List[str] = self.get_config() return config, pixel_values, labels def A__ ( self: Optional[int] ) -> int: return DeiTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowerCamelCase_ ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def A__ ( self: Optional[int] ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Any ,lowerCamelCase_: Any ) -> str: UpperCAmelCase_ : Tuple = DeiTModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : str = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self: str ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: int ,lowerCamelCase_: Dict ) -> Any: UpperCAmelCase_ : str = DeiTForMaskedImageModeling(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Tuple = model(lowerCamelCase_ ) self.parent.assertEqual( result.reconstruction.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ : Dict = 1 UpperCAmelCase_ : int = DeiTForMaskedImageModeling(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : Tuple = model(lowerCamelCase_ ) self.parent.assertEqual(result.reconstruction.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self: int ,lowerCamelCase_: Tuple ,lowerCamelCase_: str ,lowerCamelCase_: Dict ) -> Any: UpperCAmelCase_ : Any = self.type_sequence_label_size UpperCAmelCase_ : Union[str, Any] = DeiTForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Dict = model(lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase_ : Optional[int] = 1 UpperCAmelCase_ : Optional[Any] = DeiTForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : int = model(lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def A__ ( self: Dict ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : int = config_and_inputs UpperCAmelCase_ : str = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Dict = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) A__ : Optional[Any] = ( { "feature-extraction": DeiTModel, "image-classification": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) A__ : Dict = False A__ : List[Any] = False A__ : Any = False def A__ ( self: Dict ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = DeiTModelTester(self ) UpperCAmelCase_ : str = ConfigTester(self ,config_class=lowerCamelCase_ ,has_text_modality=lowerCamelCase_ ,hidden_size=37 ) def A__ ( self: str ) -> str: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def A__ ( self: str ) -> Union[str, Any]: pass def A__ ( self: Optional[int] ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = model_class(lowerCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) UpperCAmelCase_ : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ ,nn.Linear ) ) def A__ ( self: Union[str, Any] ) -> int: UpperCAmelCase_ , UpperCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : List[str] = model_class(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : str = [*signature.parameters.keys()] UpperCAmelCase_ : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase_ ) def A__ ( self: List[Any] ) -> Any: UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def A__ ( self: str ) -> Dict: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase_ ) def A__ ( self: List[str] ) -> List[str]: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) def A__ ( self: Tuple ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Tuple=False ) -> Dict: UpperCAmelCase_ : Dict = super()._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ,return_labels=lowerCamelCase_ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def A__ ( self: List[str] ) -> List[Any]: if not self.model_tester.is_training: return UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Any = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCamelCase_ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue UpperCAmelCase_ : str = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.train() UpperCAmelCase_ : List[Any] = self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ,return_labels=lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = model(**lowerCamelCase_ ).loss loss.backward() def A__ ( self: List[str] ) -> Tuple: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase_ : Optional[Any] = False UpperCAmelCase_ : Any = True for model_class in self.all_model_classes: if model_class in get_values(lowerCamelCase_ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue UpperCAmelCase_ : List[Any] = model_class(lowerCamelCase_ ) model.gradient_checkpointing_enable() model.to(lowerCamelCase_ ) model.train() UpperCAmelCase_ : Dict = self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ,return_labels=lowerCamelCase_ ) UpperCAmelCase_ : Dict = model(**lowerCamelCase_ ).loss loss.backward() def A__ ( self: Any ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Dict = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCamelCase_ ), *get_values(lowerCamelCase_ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type['title']}''' ): UpperCAmelCase_ : List[str] = problem_type["""title"""] UpperCAmelCase_ : List[Any] = problem_type["""num_labels"""] UpperCAmelCase_ : int = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.train() UpperCAmelCase_ : List[str] = self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ,return_labels=lowerCamelCase_ ) if problem_type["num_labels"] > 1: UpperCAmelCase_ : Any = inputs["""labels"""].unsqueeze(1 ).repeat(1 ,problem_type["""num_labels"""] ) UpperCAmelCase_ : Dict = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCamelCase_ ) as warning_list: UpperCAmelCase_ : Union[str, Any] = model(**lowerCamelCase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def A__ ( self: Any ) -> str: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Tuple = DeiTModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def lowerCamelCase_ ( ): '''simple docstring''' UpperCAmelCase_ : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _snake_case ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self: List[Any] ) -> int: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def A__ ( self: Dict ) -> int: UpperCAmelCase_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( lowerCamelCase_ ) UpperCAmelCase_ : List[str] = self.default_image_processor UpperCAmelCase_ : Union[str, Any] = prepare_img() UpperCAmelCase_ : List[str] = image_processor(images=lowerCamelCase_ ,return_tensors="""pt""" ).to(lowerCamelCase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : str = model(**lowerCamelCase_ ) # verify the logits UpperCAmelCase_ : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = torch.tensor([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ).to(lowerCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def A__ ( self: str ) -> Tuple: UpperCAmelCase_ : Union[str, Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" ,torch_dtype=torch.floataa ,device_map="""auto""" ) UpperCAmelCase_ : Optional[int] = self.default_image_processor UpperCAmelCase_ : Optional[int] = prepare_img() UpperCAmelCase_ : Tuple = image_processor(images=lowerCamelCase_ ,return_tensors="""pt""" ) UpperCAmelCase_ : Tuple = inputs.pixel_values.to(lowerCamelCase_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): UpperCAmelCase_ : List[Any] = model(lowerCamelCase_ )
345
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED UpperCamelCase_ = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } UpperCamelCase_ = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowerCamelCase_ ( ): '''simple docstring''' UpperCAmelCase_ : int = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) UpperCAmelCase_ : Dict = bs[:] UpperCAmelCase_ : Any = 0 for b in range(2**8 ): if b not in bs: bs.append(_a ) cs.append(2**8 + n ) n += 1 UpperCAmelCase_ : Any = [chr(_a ) for n in cs] return dict(zip(_a , _a ) ) def lowerCamelCase_ ( _a : List[str] ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = set() UpperCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase_ : Optional[int] = char return pairs class _snake_case ( __snake_case ): '''simple docstring''' A__ : str = VOCAB_FILES_NAMES A__ : List[str] = PRETRAINED_VOCAB_FILES_MAP A__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[int] = ["input_ids", "attention_mask"] def __init__( self: Union[str, Any] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Any ,lowerCamelCase_: Union[str, Any]="replace" ,lowerCamelCase_: Optional[Any]="<s>" ,lowerCamelCase_: List[Any]="</s>" ,lowerCamelCase_: List[str]="</s>" ,lowerCamelCase_: int="<s>" ,lowerCamelCase_: int="<unk>" ,lowerCamelCase_: str="<pad>" ,lowerCamelCase_: Optional[Any]="<mask>" ,lowerCamelCase_: List[str]=False ,**lowerCamelCase_: Tuple ,) -> Any: UpperCAmelCase_ : Union[str, Any] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else bos_token UpperCAmelCase_ : int = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else eos_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else sep_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else cls_token UpperCAmelCase_ : Optional[Any] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else unk_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ : str = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else mask_token super().__init__( errors=lowerCamelCase_ ,bos_token=lowerCamelCase_ ,eos_token=lowerCamelCase_ ,unk_token=lowerCamelCase_ ,sep_token=lowerCamelCase_ ,cls_token=lowerCamelCase_ ,pad_token=lowerCamelCase_ ,mask_token=lowerCamelCase_ ,add_prefix_space=lowerCamelCase_ ,**lowerCamelCase_ ,) with open(lowerCamelCase_ ,encoding="""utf-8""" ) as vocab_handle: UpperCAmelCase_ : Union[str, Any] = json.load(lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} UpperCAmelCase_ : Any = errors # how to handle errors in decoding UpperCAmelCase_ : int = bytes_to_unicode() UpperCAmelCase_ : Dict = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase_ ,encoding="""utf-8""" ) as merges_handle: UpperCAmelCase_ : Any = merges_handle.read().split("""\n""" )[1:-1] UpperCAmelCase_ : int = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase_ : Union[str, Any] = dict(zip(lowerCamelCase_ ,range(len(lowerCamelCase_ ) ) ) ) UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : Optional[int] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase_ : int = re.compile(R"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def A__ ( self: List[str] ) -> List[str]: return len(self.encoder ) def A__ ( self: Any ) -> Union[str, Any]: return dict(self.encoder ,**self.added_tokens_encoder ) def A__ ( self: Tuple ,lowerCamelCase_: Dict ) -> Optional[Any]: if token in self.cache: return self.cache[token] UpperCAmelCase_ : Union[str, Any] = tuple(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = get_pairs(lowerCamelCase_ ) if not pairs: return token while True: UpperCAmelCase_ : Union[str, Any] = min(lowerCamelCase_ ,key=lambda lowerCamelCase_ : self.bpe_ranks.get(lowerCamelCase_ ,float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase_ , UpperCAmelCase_ : Any = bigram UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : List[str] = 0 while i < len(lowerCamelCase_ ): try: UpperCAmelCase_ : str = word.index(lowerCamelCase_ ,lowerCamelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase_ : Union[str, Any] = j if word[i] == first and i < len(lowerCamelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase_ : List[str] = tuple(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = new_word if len(lowerCamelCase_ ) == 1: break else: UpperCAmelCase_ : List[str] = get_pairs(lowerCamelCase_ ) UpperCAmelCase_ : int = """ """.join(lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = word return word def A__ ( self: Union[str, Any] ,lowerCamelCase_: Tuple ) -> List[str]: UpperCAmelCase_ : str = [] for token in re.findall(self.pat ,lowerCamelCase_ ): UpperCAmelCase_ : List[Any] = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase_ ).split(""" """ ) ) return bpe_tokens def A__ ( self: List[Any] ,lowerCamelCase_: Optional[Any] ) -> Optional[int]: return self.encoder.get(lowerCamelCase_ ,self.encoder.get(self.unk_token ) ) def A__ ( self: List[str] ,lowerCamelCase_: str ) -> Optional[Any]: return self.decoder.get(lowerCamelCase_ ) def A__ ( self: List[str] ,lowerCamelCase_: List[str] ) -> List[Any]: UpperCAmelCase_ : str = """""".join(lowerCamelCase_ ) UpperCAmelCase_ : int = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" ,errors=self.errors ) return text def A__ ( self: Optional[Any] ,lowerCamelCase_: str ,lowerCamelCase_: Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : List[Any] = os.path.join( lowerCamelCase_ ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase_ : List[str] = os.path.join( lowerCamelCase_ ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCamelCase_ ,"""w""" ,encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase_ ,ensure_ascii=lowerCamelCase_ ) + """\n""" ) UpperCAmelCase_ : str = 0 with open(lowerCamelCase_ ,"""w""" ,encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase_ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) UpperCAmelCase_ : Tuple = token_index writer.write(""" """.join(lowerCamelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file def A__ ( self: str ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ : int = [self.cls_token_id] UpperCAmelCase_ : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self: Union[str, Any] ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ,lowerCamelCase_: bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ ,token_ids_a=lowerCamelCase_ ,already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase_ )) + [1] return [1] + ([0] * len(lowerCamelCase_ )) + [1, 1] + ([0] * len(lowerCamelCase_ )) + [1] def A__ ( self: str ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ) -> List[int]: UpperCAmelCase_ : Optional[Any] = [self.sep_token_id] UpperCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self: Optional[Any] ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: str=False ,**lowerCamelCase_: List[str] ) -> Optional[int]: UpperCAmelCase_ : Optional[int] = kwargs.pop("""add_prefix_space""" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase_ ) > 0 and not text[0].isspace()): UpperCAmelCase_ : Dict = """ """ + text return (text, kwargs) def A__ ( self: List[str] ,lowerCamelCase_: Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCamelCase_: Optional[int] = None ,lowerCamelCase_: PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase_: Optional[int] = None ,lowerCamelCase_: Optional[bool] = None ,) -> dict: UpperCAmelCase_ : Optional[int] = super()._pad( encoded_inputs=lowerCamelCase_ ,max_length=lowerCamelCase_ ,padding_strategy=lowerCamelCase_ ,pad_to_multiple_of=lowerCamelCase_ ,return_attention_mask=lowerCamelCase_ ,) # Load from model defaults if return_attention_mask is None: UpperCAmelCase_ : str = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: UpperCAmelCase_ : str = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. UpperCAmelCase_ : List[Any] = len(encoded_inputs["""global_attention_mask"""] ) != len(lowerCamelCase_ ) if needs_to_be_padded: UpperCAmelCase_ : Dict = len(lowerCamelCase_ ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` UpperCAmelCase_ : str = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": UpperCAmelCase_ : List[str] = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
345
1
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class UpperCamelCase ( enum.Enum ): UpperCAmelCase : Union[str, Any] = 0 UpperCAmelCase : Optional[Any] = 1 UpperCAmelCase : Any = 2 @add_end_docstrings(lowercase ) class UpperCamelCase ( lowercase ): UpperCAmelCase : Optional[Any] = """ In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas's young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> """ def __init__(self : Dict , *_A : int , **_A : Dict) -> Dict: super().__init__(*_A , **_A) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == 'tf' else MODEL_FOR_CAUSAL_LM_MAPPING) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. __snake_case : Any = None if self.model.config.prefix is not None: __snake_case : str = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. __snake_case : Optional[Any] = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. __snake_case , __snake_case , __snake_case : Optional[int] = self._sanitize_parameters(prefix=_A , **self._forward_params) __snake_case : Any = {**self._preprocess_params, **preprocess_params} __snake_case : List[Any] = {**self._forward_params, **forward_params} def _lowercase (self : Union[str, Any] , _A : List[str]=None , _A : Optional[Any]=None , _A : Tuple=None , _A : Any=None , _A : List[Any]=None , _A : Tuple=None , _A : str=None , _A : str=None , **_A : int , ) -> List[str]: __snake_case : Union[str, Any] = {} if prefix is not None: __snake_case : List[Any] = prefix if prefix: __snake_case : Optional[Any] = self.tokenizer( _A , padding=_A , add_special_tokens=_A , return_tensors=self.framework) __snake_case : List[Any] = prefix_inputs['input_ids'].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f"{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected" ' [None, \'hole\']') __snake_case : Dict = handle_long_generation preprocess_params.update(_A) __snake_case : List[Any] = generate_kwargs __snake_case : Optional[Any] = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError('`return_text` is mutually exclusive with `return_full_text`') if return_tensors is not None: raise ValueError('`return_full_text` is mutually exclusive with `return_tensors`') __snake_case : str = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError('`return_text` is mutually exclusive with `return_tensors`') __snake_case : Tuple = ReturnType.TENSORS if return_type is not None: __snake_case : Any = return_type if clean_up_tokenization_spaces is not None: __snake_case : List[Any] = clean_up_tokenization_spaces if stop_sequence is not None: __snake_case : Optional[Any] = self.tokenizer.encode(_A , add_special_tokens=_A) if len(_A) > 1: warnings.warn( 'Stopping on a multiple token sequence is not yet supported on transformers. The first token of' ' the stop sequence will be used as the stop sequence string in the interim.') __snake_case : Dict = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _lowercase (self : str , *_A : Dict , **_A : List[str]) -> Any: # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({'add_space_before_punct_symbol': True}) return super()._parse_and_tokenize(*_A , **_A) def __call__(self : List[str] , _A : Optional[Any] , **_A : str) -> Any: return super().__call__(_A , **_A) def _lowercase (self : Optional[int] , _A : int , _A : str="" , _A : str=None , **_A : List[Any]) -> Optional[int]: __snake_case : List[str] = self.tokenizer( prefix + prompt_text , padding=_A , add_special_tokens=_A , return_tensors=self.framework) __snake_case : Any = prompt_text if handle_long_generation == "hole": __snake_case : Tuple = inputs['input_ids'].shape[-1] if "max_new_tokens" in generate_kwargs: __snake_case : Tuple = generate_kwargs['max_new_tokens'] else: __snake_case : int = generate_kwargs.get('max_length' , self.model.config.max_length) - cur_len if new_tokens < 0: raise ValueError('We cannot infer how many new tokens are expected') if cur_len + new_tokens > self.tokenizer.model_max_length: __snake_case : Union[str, Any] = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( 'We cannot use `hole` to handle this generation the number of desired tokens exceeds the' ' models max length') __snake_case : Union[str, Any] = inputs['input_ids'][:, -keep_length:] if "attention_mask" in inputs: __snake_case : int = inputs['attention_mask'][:, -keep_length:] return inputs def _lowercase (self : List[Any] , _A : List[str] , **_A : Optional[Any]) -> str: __snake_case : Union[str, Any] = model_inputs['input_ids'] __snake_case : Tuple = model_inputs.get('attention_mask' , _A) # Allow empty prompts if input_ids.shape[1] == 0: __snake_case : Optional[int] = None __snake_case : Dict = None __snake_case : Union[str, Any] = 1 else: __snake_case : Optional[int] = input_ids.shape[0] __snake_case : Union[str, Any] = model_inputs.pop('prompt_text') # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. __snake_case : int = generate_kwargs.pop('prefix_length' , 0) if prefix_length > 0: __snake_case : Optional[int] = 'max_new_tokens' in generate_kwargs or ( 'generation_config' in generate_kwargs and generate_kwargs['generation_config'].max_new_tokens is not None ) if not has_max_new_tokens: __snake_case : int = generate_kwargs.get('max_length') or self.model.config.max_length generate_kwargs["max_length"] += prefix_length __snake_case : Union[str, Any] = 'min_new_tokens' in generate_kwargs or ( 'generation_config' in generate_kwargs and generate_kwargs['generation_config'].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL __snake_case : Dict = self.model.generate(input_ids=_A , attention_mask=_A , **_A) __snake_case : str = generated_sequence.shape[0] if self.framework == "pt": __snake_case : Any = generated_sequence.reshape(_A , out_b // in_b , *generated_sequence.shape[1:]) elif self.framework == "tf": __snake_case : Tuple = tf.reshape(_A , (in_b, out_b // in_b, *generated_sequence.shape[1:])) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def _lowercase (self : str , _A : List[str] , _A : Dict=ReturnType.FULL_TEXT , _A : Any=True) -> Tuple: __snake_case : Dict = model_outputs['generated_sequence'][0] __snake_case : List[str] = model_outputs['input_ids'] __snake_case : Optional[Any] = model_outputs['prompt_text'] __snake_case : Any = generated_sequence.numpy().tolist() __snake_case : Tuple = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: __snake_case : Tuple = {'generated_token_ids': sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text __snake_case : Tuple = self.tokenizer.decode( _A , skip_special_tokens=_A , clean_up_tokenization_spaces=_A , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: __snake_case : int = 0 else: __snake_case : Optional[int] = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_A , clean_up_tokenization_spaces=_A , )) if return_type == ReturnType.FULL_TEXT: __snake_case : Union[str, Any] = prompt_text + text[prompt_length:] else: __snake_case : int = text[prompt_length:] __snake_case : List[Any] = {'generated_text': all_text} records.append(_A) return records
95
"""simple docstring""" from __future__ import annotations import os from typing import Any import requests _a : int= "https://api.github.com" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _a : Dict= BASE_URL + "/user" # https://github.com/settings/tokens _a : Union[str, Any]= os.environ.get("USER_TOKEN", "") def __UpperCAmelCase ( UpperCAmelCase_ : str ) -> dict[Any, Any]: '''simple docstring''' __snake_case : Tuple = { 'Authorization': F"token {auth_token}", 'Accept': 'application/vnd.github.v3+json', } return requests.get(UpperCAmelCase_ , headers=UpperCAmelCase_ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(f'''{key}: {value}''') else: raise ValueError("'USER_TOKEN' field cannot be empty.")
95
1
'''simple docstring''' from math import factorial def _UpperCamelCase ( __A = 20 ) -> int: '''simple docstring''' UpperCamelCase__ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... UpperCamelCase__ = n // 2 return int(factorial(__A ) / (factorial(__A ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(2_0)) else: try: a__ : List[Any] = int(sys.argv[1]) print(solution(n)) except ValueError: print('Invalid entry - please enter a number.')
80
'''simple docstring''' # Function to print upper half of diamond (pyramid) def __snake_case( _lowerCAmelCase ) -> Any: for i in range(0 , _lowerCAmelCase ): for _ in range(0 , n - i - 1 ): # printing spaces print(""" """ , end="""""" ) for _ in range(0 , i + 1 ): # printing stars print("""* """ , end="""""" ) print() def __snake_case( _lowerCAmelCase ) -> List[str]: for i in range(_lowerCAmelCase , 0 , -1 ): for _ in range(_lowerCAmelCase , 0 , -1 ): # printing stars print("""* """ , end="""""" ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(""" """ , end="""""" ) def __snake_case( _lowerCAmelCase ) -> List[Any]: if n <= 0: print(""" ... .... nothing printing :(""" ) return floyd(_lowerCAmelCase ) # upper half reverse_floyd(_lowerCAmelCase ) # lower half if __name__ == "__main__": print(R"| /\ | |- | |- |--| |\ /| |-") print(R"|/ \| |- |_ |_ |__| | \/ | |_") __a = 1 while K: __a = int(input("enter the number and , and see the magic : ")) print() pretty_print(user_number) __a = int(input("press 0 to exit... and 1 to continue...")) print("Good Bye...")
35
0
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCAmelCase_ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> List[str]: UpperCamelCase__ : int = argparse.ArgumentParser( description='''Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).''' ) parser.add_argument('''--file_path''' , type=_a , default='''data/dump.txt''' , help='''The path to the data.''' ) parser.add_argument('''--tokenizer_type''' , type=_a , default='''bert''' , choices=['''bert''', '''roberta''', '''gpt2'''] ) parser.add_argument('''--tokenizer_name''' , type=_a , default='''bert-base-uncased''' , help='''The tokenizer to use.''' ) parser.add_argument('''--dump_file''' , type=_a , default='''data/dump''' , help='''The dump file prefix.''' ) UpperCamelCase__ : Tuple = parser.parse_args() logger.info(f"Loading Tokenizer ({args.tokenizer_name})" ) if args.tokenizer_type == "bert": UpperCamelCase__ : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCamelCase__ : Dict = tokenizer.special_tokens_map['''cls_token'''] # `[CLS]` UpperCamelCase__ : Tuple = tokenizer.special_tokens_map['''sep_token'''] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCamelCase__ : List[str] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCamelCase__ : Optional[int] = tokenizer.special_tokens_map['''cls_token'''] # `<s>` UpperCamelCase__ : int = tokenizer.special_tokens_map['''sep_token'''] # `</s>` elif args.tokenizer_type == "gpt2": UpperCamelCase__ : List[str] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCamelCase__ : Dict = tokenizer.special_tokens_map['''bos_token'''] # `<|endoftext|>` UpperCamelCase__ : Optional[int] = tokenizer.special_tokens_map['''eos_token'''] # `<|endoftext|>` logger.info(f"Loading text from {args.file_path}" ) with open(args.file_path , '''r''' , encoding='''utf8''' ) as fp: UpperCamelCase__ : Union[str, Any] = fp.readlines() logger.info('''Start encoding''' ) logger.info(f"{len(_a )} examples to process." ) UpperCamelCase__ : Optional[Any] = [] UpperCamelCase__ : Dict = 0 UpperCamelCase__ : int = 1_0000 UpperCamelCase__ : Tuple = time.time() for text in data: UpperCamelCase__ : Tuple = f"{bos} {text.strip()} {sep}" UpperCamelCase__ : int = tokenizer.encode(_a , add_special_tokens=_a ) rslt.append(_a ) iter += 1 if iter % interval == 0: UpperCamelCase__ : Any = time.time() logger.info(f"{iter} examples processed. - {(end-start):.2f}s/{interval}expl" ) UpperCamelCase__ : Optional[Any] = time.time() logger.info('''Finished binarization''' ) logger.info(f"{len(_a )} examples processed." ) UpperCamelCase__ : int = f"{args.dump_file}.{args.tokenizer_name}.pickle" UpperCamelCase__ : Optional[Any] = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCamelCase__ : List[Any] = [np.uintaa(_a ) for d in rslt] else: UpperCamelCase__ : int = [np.intaa(_a ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"Dump to {dp_file}" ) with open(_a , '''wb''' ) as handle: pickle.dump(rslt_ , _a , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
361
from manim import * class lowercase__ ( __lowerCamelCase ): '''simple docstring''' def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : int = Rectangle(height=0.5, width=0.5 ) UpperCamelCase__ : Optional[int] = Rectangle(height=0.46, width=0.46 ).set_stroke(width=0 ) UpperCamelCase__ : Dict = [mem.copy() for i in range(6 )] UpperCamelCase__ : Any = [mem.copy() for i in range(6 )] UpperCamelCase__ : int = VGroup(*__magic_name__ ).arrange(__magic_name__, buff=0 ) UpperCamelCase__ : Tuple = VGroup(*__magic_name__ ).arrange(__magic_name__, buff=0 ) UpperCamelCase__ : int = VGroup(__magic_name__, __magic_name__ ).arrange(__magic_name__, buff=0 ) UpperCamelCase__ : Optional[int] = Text('''CPU''', font_size=24 ) UpperCamelCase__ : Any = Group(__magic_name__, __magic_name__ ).arrange(__magic_name__, buff=0.5, aligned_edge=__magic_name__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__magic_name__ ) UpperCamelCase__ : Any = [mem.copy() for i in range(1 )] UpperCamelCase__ : Optional[int] = VGroup(*__magic_name__ ).arrange(__magic_name__, buff=0 ) UpperCamelCase__ : Union[str, Any] = Text('''GPU''', font_size=24 ) UpperCamelCase__ : List[Any] = Group(__magic_name__, __magic_name__ ).arrange(__magic_name__, buff=0.5, aligned_edge=__magic_name__ ) gpu.align_to(__magic_name__, __magic_name__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(__magic_name__ ) UpperCamelCase__ : str = [mem.copy() for i in range(6 )] UpperCamelCase__ : Optional[int] = VGroup(*__magic_name__ ).arrange(__magic_name__, buff=0 ) UpperCamelCase__ : Optional[int] = Text('''Model''', font_size=24 ) UpperCamelCase__ : int = Group(__magic_name__, __magic_name__ ).arrange(__magic_name__, buff=0.5, aligned_edge=__magic_name__ ) model.move_to([3, -1.0, 0] ) self.play( Create(__magic_name__, run_time=1 ), Create(__magic_name__, run_time=1 ), Create(__magic_name__, run_time=1 ), ) UpperCamelCase__ : Optional[int] = MarkupText( f"First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.", font_size=24, ) UpperCamelCase__ : List[str] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCamelCase__ : Union[str, Any] = MarkupText( f"<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model", font_size=18, ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(__magic_name__, run_time=2.5 ), Write(__magic_name__ ), Write(__magic_name__ ) ) self.add(__magic_name__ ) UpperCamelCase__ : Dict = [] UpperCamelCase__ : Any = [] UpperCamelCase__ : int = [] for i, rect in enumerate(__magic_name__ ): UpperCamelCase__ : Union[str, Any] = Rectangle(height=0.46, width=0.46 ).set_stroke(width=0.0 ).set_fill(__magic_name__, opacity=0.7 ) cpu_target.move_to(__magic_name__ ) cpu_target.generate_target() UpperCamelCase__ : Tuple = 0.46 / 4 UpperCamelCase__ : Optional[Any] = 0.46 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ), buff=0.02, direction=__magic_name__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target, direction=__magic_name__, buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target, direction=__magic_name__, buff=0.0 ) cpu_targs.append(__magic_name__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(__magic_name__ ) ) second_animations.append(MoveToTarget(__magic_name__, run_time=1.5 ) ) self.play(*__magic_name__ ) self.play(*__magic_name__ ) self.wait()
247
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable _snake_case = {'''configuration_gpt_neox''': ['''GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXConfig''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['''GPTNeoXTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ '''GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXForCausalLM''', '''GPTNeoXForQuestionAnswering''', '''GPTNeoXForSequenceClassification''', '''GPTNeoXForTokenClassification''', '''GPTNeoXLayer''', '''GPTNeoXModel''', '''GPTNeoXPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
157
"""simple docstring""" from functools import lru_cache @lru_cache def UpperCamelCase__ ( lowercase__ : int ): if num < 0: raise ValueError("Number should not be negative." ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
148
0
import argparse import os import re import packaging.version _A = """examples/""" _A = { """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } _A = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } _A = """README.md""" def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Optional[int]: with open(lowerCAmelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase__ : List[Any] = f.read() UpperCAmelCase__ : Tuple = REPLACE_PATTERNS[pattern] UpperCAmelCase__ : Dict = replace.replace("""VERSION""" , lowerCAmelCase ) UpperCAmelCase__ : Any = re_pattern.sub(lowerCAmelCase , lowerCAmelCase ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(lowerCAmelCase ) def a__ ( lowerCAmelCase ) -> List[str]: for folder, directories, fnames in os.walk(lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase , pattern="""examples""" ) def a__ ( lowerCAmelCase , lowerCAmelCase=False ) -> Optional[int]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if not patch: update_version_in_examples(lowerCAmelCase ) def a__ ( ) -> int: UpperCAmelCase__ : Optional[int] = """🤗 Transformers currently provides the following architectures""" UpperCAmelCase__ : Union[str, Any] = """1. Want to contribute a new model?""" with open(lowerCAmelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase__ : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase__ : Optional[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase__ : Tuple = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): UpperCAmelCase__ : Tuple = lines[index].replace( """https://huggingface.co/docs/transformers/main/model_doc""" , """https://huggingface.co/docs/transformers/model_doc""" , ) index += 1 with open(lowerCAmelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase ) def a__ ( ) -> Tuple: with open(REPLACE_FILES["""init"""] , """r""" ) as f: UpperCAmelCase__ : Union[str, Any] = f.read() UpperCAmelCase__ : Union[str, Any] = REPLACE_PATTERNS["""init"""][0].search(lowerCAmelCase ).groups()[0] return packaging.version.parse(lowerCAmelCase ) def a__ ( lowerCAmelCase=False ) -> Union[str, Any]: UpperCAmelCase__ : str = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: UpperCAmelCase__ : Union[str, Any] = default_version.base_version elif patch: UpperCAmelCase__ : Optional[int] = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase__ : Dict = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase__ : Any = input(F"""Which version are you releasing? [{default_version}]""" ) if len(lowerCAmelCase ) == 0: UpperCAmelCase__ : Optional[Any] = default_version print(F"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase , patch=lowerCAmelCase ) if not patch: print("""Cleaning main README, don't forget to run `make fix-copies`.""" ) clean_main_ref_in_model_list() def a__ ( ) -> Any: UpperCAmelCase__ : int = get_version() UpperCAmelCase__ : Dict = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase__ : int = current_version.base_version # Check with the user we got that right. UpperCAmelCase__ : Dict = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(lowerCAmelCase ) == 0: UpperCAmelCase__ : int = dev_version print(F"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase ) print("""Cleaning main README, don't forget to run `make fix-copies`.""" ) clean_main_ref_in_model_list() if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") _A = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
362
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder _A = """__DUMMY_TRANSFORMERS_USER__""" _A = """Dummy User""" _A = """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""" _A = """https://hub-ci.huggingface.co""" _A = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" _A = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" _A = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def a__ ( lowerCAmelCase ) -> Union[str, Any]: monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""" , lowerCAmelCase ) @pytest.fixture def a__ ( lowerCAmelCase ) -> List[Any]: monkeypatch.setattr("""datasets.config.HF_ENDPOINT""" , lowerCAmelCase ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""" , lowerCAmelCase ) @pytest.fixture def a__ ( lowerCAmelCase ) -> List[Any]: monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""" , lowerCAmelCase ) @pytest.fixture def a__ ( lowerCAmelCase , lowerCAmelCase ) -> str: HfFolder.save_token(lowerCAmelCase ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def a__ ( ) -> List[str]: return HfApi(endpoint=lowerCAmelCase ) @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase ) -> Union[str, Any]: UpperCAmelCase__ : List[str] = HfFolder.get_token() HfFolder.save_token(lowerCAmelCase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(lowerCAmelCase ) @pytest.fixture def a__ ( lowerCAmelCase ) -> List[str]: def _cleanup_repo(lowerCAmelCase ): hf_api.delete_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def a__ ( lowerCAmelCase ) -> Optional[Any]: @contextmanager def _temporary_repo(lowerCAmelCase ): try: yield repo_id finally: cleanup_repo(lowerCAmelCase ) return _temporary_repo @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Union[str, Any]: UpperCAmelCase__ : str = F"""repo_txt_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : List[str] = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" , private=lowerCAmelCase ) hf_api.upload_file( token=lowerCAmelCase , path_or_fileobj=str(lowerCAmelCase ) , path_in_repo="""data/text_data.txt""" , repo_id=lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> List[Any]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> int: UpperCAmelCase__ : List[Any] = F"""repo_zipped_txt_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : Any = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" , private=lowerCAmelCase ) hf_api.upload_file( token=lowerCAmelCase , path_or_fileobj=str(lowerCAmelCase ) , path_in_repo="""data.zip""" , repo_id=lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Dict: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Tuple: UpperCAmelCase__ : Union[str, Any] = F"""repo_zipped_img_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : Optional[int] = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" , private=lowerCAmelCase ) hf_api.upload_file( token=lowerCAmelCase , path_or_fileobj=str(lowerCAmelCase ) , path_in_repo="""data.zip""" , repo_id=lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(lowerCAmelCase , token=lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Optional[Any]: return hf_private_dataset_repo_zipped_img_data_
166
0
"""simple docstring""" from collections import deque from math import floor from random import random from time import time class lowerCAmelCase_ : """simple docstring""" def __init__(self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = {} def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 ) -> str: """simple docstring""" if self.graph.get(SCREAMING_SNAKE_CASE__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: SCREAMING_SNAKE_CASE__ : Any = [[w, v]] if not self.graph.get(SCREAMING_SNAKE_CASE__ ): SCREAMING_SNAKE_CASE__ : List[str] = [] def __magic_name__ (self ) -> List[Any]: """simple docstring""" return list(self.graph ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> Dict: """simple docstring""" if s == d: return [] SCREAMING_SNAKE_CASE__ : int = [] SCREAMING_SNAKE_CASE__ : Tuple = [] if s == -2: SCREAMING_SNAKE_CASE__ : Any = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(SCREAMING_SNAKE_CASE__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : List[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Optional[int] = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : Tuple = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return visited def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-1 ) -> int: """simple docstring""" if c == -1: SCREAMING_SNAKE_CASE__ : int = floor(random() * 1_00_00 ) + 10 for i in range(SCREAMING_SNAKE_CASE__ ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): SCREAMING_SNAKE_CASE__ : Tuple = floor(random() * c ) + 1 if n != i: self.add_pair(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = deque() SCREAMING_SNAKE_CASE__ : str = [] if s == -2: SCREAMING_SNAKE_CASE__ : str = list(self.graph )[0] d.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) while d: SCREAMING_SNAKE_CASE__ : Dict = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" return len(self.graph[u] ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = [] SCREAMING_SNAKE_CASE__ : Dict = [] if s == -2: SCREAMING_SNAKE_CASE__ : Any = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[Any] = s SCREAMING_SNAKE_CASE__ : Optional[Any] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : Tuple = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Optional[int] = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : Optional[int] = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return sorted_nodes def __magic_name__ (self ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = [] SCREAMING_SNAKE_CASE__ : Tuple = [] SCREAMING_SNAKE_CASE__ : List[Any] = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = -2 SCREAMING_SNAKE_CASE__ : int = [] SCREAMING_SNAKE_CASE__ : int = s SCREAMING_SNAKE_CASE__ : Optional[Any] = False SCREAMING_SNAKE_CASE__ : Union[str, Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : str = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ : Tuple = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : List[str] = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ : Tuple = True if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : List[str] = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = s SCREAMING_SNAKE_CASE__ : Optional[int] = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return list(SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = [] SCREAMING_SNAKE_CASE__ : Dict = [] SCREAMING_SNAKE_CASE__ : Optional[Any] = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = -2 SCREAMING_SNAKE_CASE__ : Optional[Any] = [] SCREAMING_SNAKE_CASE__ : Optional[int] = s SCREAMING_SNAKE_CASE__ : Tuple = False SCREAMING_SNAKE_CASE__ : Tuple = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : int = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ : Any = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ : List[Any] = True if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Tuple = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : Optional[int] = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[int] = s SCREAMING_SNAKE_CASE__ : int = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return False def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = time() self.dfs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = time() return end - begin def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = time() self.bfs(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = time() return end - begin class lowerCAmelCase_ : """simple docstring""" def __init__(self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = {} def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 ) -> int: """simple docstring""" if self.graph.get(SCREAMING_SNAKE_CASE__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist SCREAMING_SNAKE_CASE__ : Any = [[w, v]] # add the other way if self.graph.get(SCREAMING_SNAKE_CASE__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist SCREAMING_SNAKE_CASE__ : Optional[Any] = [[w, u]] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Optional[int]: """simple docstring""" if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(SCREAMING_SNAKE_CASE__ ) # the other way round if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> Optional[int]: """simple docstring""" if s == d: return [] SCREAMING_SNAKE_CASE__ : Any = [] SCREAMING_SNAKE_CASE__ : str = [] if s == -2: SCREAMING_SNAKE_CASE__ : Optional[Any] = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(SCREAMING_SNAKE_CASE__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : Dict = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : List[str] = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return visited def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-1 ) -> Tuple: """simple docstring""" if c == -1: SCREAMING_SNAKE_CASE__ : Dict = floor(random() * 1_00_00 ) + 10 for i in range(SCREAMING_SNAKE_CASE__ ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): SCREAMING_SNAKE_CASE__ : str = floor(random() * c ) + 1 if n != i: self.add_pair(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = deque() SCREAMING_SNAKE_CASE__ : int = [] if s == -2: SCREAMING_SNAKE_CASE__ : str = list(self.graph )[0] d.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) while d: SCREAMING_SNAKE_CASE__ : List[Any] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" return len(self.graph[u] ) def __magic_name__ (self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = [] SCREAMING_SNAKE_CASE__ : int = [] SCREAMING_SNAKE_CASE__ : Union[str, Any] = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = -2 SCREAMING_SNAKE_CASE__ : Tuple = [] SCREAMING_SNAKE_CASE__ : Optional[Any] = s SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : int = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ : Any = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ : Union[str, Any] = True if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : int = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[str] = s SCREAMING_SNAKE_CASE__ : Any = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return list(SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = [] SCREAMING_SNAKE_CASE__ : Dict = [] SCREAMING_SNAKE_CASE__ : Any = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : str = -2 SCREAMING_SNAKE_CASE__ : Dict = [] SCREAMING_SNAKE_CASE__ : Optional[int] = s SCREAMING_SNAKE_CASE__ : List[str] = False SCREAMING_SNAKE_CASE__ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ : Tuple = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ : str = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ : Optional[Any] = True if len(SCREAMING_SNAKE_CASE__ ) != 0: SCREAMING_SNAKE_CASE__ : Dict = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: SCREAMING_SNAKE_CASE__ : List[Any] = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[int] = s SCREAMING_SNAKE_CASE__ : Tuple = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return False def __magic_name__ (self ) -> Optional[int]: """simple docstring""" return list(self.graph ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = time() self.dfs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Any = time() return end - begin def __magic_name__ (self , SCREAMING_SNAKE_CASE__=-2 ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = time() self.bfs(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Any = time() return end - begin
25
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __snake_case ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = """facebook/bart-large-mnli""" _lowerCamelCase = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) _lowerCamelCase = """text_classifier""" _lowerCamelCase = AutoTokenizer _lowerCamelCase = AutoModelForSequenceClassification _lowerCamelCase = ["""text""", ["""text"""]] _lowerCamelCase = ["""text"""] def UpperCamelCase__( self ): '''simple docstring''' super().setup() __A : List[str] = self.model.config __A : int = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): __A : List[str] = int(__lowerCamelCase ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' __A : Union[str, Any] = labels return self.pre_processor( [text] * len(__lowerCamelCase ) , [F"""This example is {label}""" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : List[Any] = outputs.logits __A : List[str] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
179
0
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[int] , lowercase_ : Tuple , lowercase_ : Tuple=13 , lowercase_ : Any=7 , lowercase_ : Union[str, Any]=True , lowercase_ : Optional[int]=True , lowercase_ : Optional[int]=False , lowercase_ : Tuple=True , lowercase_ : Tuple=99 , lowercase_ : Union[str, Any]=32 , lowercase_ : Optional[Any]=5 , lowercase_ : Optional[Any]=4 , lowercase_ : Optional[int]=37 , lowercase_ : Optional[Any]="gelu" , lowercase_ : Tuple=0.1 , lowercase_ : Dict=0.1 , lowercase_ : Union[str, Any]=512 , lowercase_ : Tuple=16 , lowercase_ : List[str]=2 , lowercase_ : Union[str, Any]=0.02 , lowercase_ : Optional[Any]=3 , lowercase_ : Union[str, Any]=4 , lowercase_ : Union[str, Any]=None , ) -> Tuple: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self : Any) -> int: """simple docstring""" return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : Tuple) -> Optional[int]: """simple docstring""" _UpperCamelCase = BioGptModel(config=snake_case__) model.to(snake_case__) model.eval() _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__) _UpperCamelCase = model(snake_case__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : str , lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Any , lowercase_ : Dict , lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[str] , ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = BioGptForCausalLM(config=snake_case__) model.to(snake_case__) model.eval() _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] , *lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = BioGptModel(config=snake_case__) model.to(snake_case__) model.eval() # create attention mask _UpperCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case__) _UpperCamelCase = self.seq_length // 2 _UpperCamelCase = 0 # first forward pass _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__).to_tuple() # create hypothetical next token and extent to next_input_ids _UpperCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size) # change a random masked slice from input_ids _UpperCamelCase = ids_tensor((1,) , snake_case__).item() + 1 _UpperCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size).squeeze(-1) _UpperCamelCase = random_other_next_tokens # append to next input_ids and attn_mask _UpperCamelCase = torch.cat([input_ids, next_tokens] , dim=-1) _UpperCamelCase = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case__)] , dim=1 , ) # get two different outputs _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__)['''last_hidden_state'''] _UpperCamelCase = model(snake_case__ , past_key_values=snake_case__ , attention_mask=snake_case__)['''last_hidden_state'''] # select random slice _UpperCamelCase = ids_tensor((1,) , output_from_past.shape[-1]).item() _UpperCamelCase = output_from_no_past[:, -1, random_slice_idx].detach() _UpperCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-3)) def __UpperCAmelCase ( self : str , lowercase_ : str , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : str , *lowercase_ : Tuple) -> int: """simple docstring""" _UpperCamelCase = BioGptModel(config=snake_case__).to(snake_case__).eval() _UpperCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case__) # first forward pass _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , use_cache=snake_case__) _UpperCamelCase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids _UpperCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size) _UpperCamelCase = ids_tensor((self.batch_size, 3) , 2) # append to next input_ids and _UpperCamelCase = torch.cat([input_ids, next_tokens] , dim=-1) _UpperCamelCase = torch.cat([attention_mask, next_attn_mask] , dim=-1) _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__)['''last_hidden_state'''] _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , past_key_values=snake_case__)[ '''last_hidden_state''' ] # select random slice _UpperCamelCase = ids_tensor((1,) , output_from_past.shape[-1]).item() _UpperCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _UpperCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-3)) def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[Any] , lowercase_ : Any , lowercase_ : Dict , lowercase_ : List[Any] , lowercase_ : Optional[int] , *lowercase_ : Dict , lowercase_ : int=False) -> Dict: """simple docstring""" _UpperCamelCase = BioGptForCausalLM(snake_case__) model.to(snake_case__) if gradient_checkpointing: model.gradient_checkpointing_enable() _UpperCamelCase = model(snake_case__ , labels=snake_case__) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) result.loss.backward() def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[Any] , *lowercase_ : Any) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = BioGptModel(snake_case__) _UpperCamelCase = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key]) - model_std) , 0.0_01) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key]) - 0.0) , 0.01) def __UpperCAmelCase ( self : Any , lowercase_ : int , lowercase_ : int , lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] , *lowercase_ : Tuple) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = BioGptForTokenClassification(snake_case__) model.to(snake_case__) model.eval() _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def __UpperCAmelCase ( self : Any) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() ( _UpperCamelCase ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, unittest.TestCase ): '''simple docstring''' __A = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) __A = (BioGptForCausalLM,) if is_torch_available() else () __A = ( { """feature-extraction""": BioGptModel, """text-classification""": BioGptForSequenceClassification, """text-generation""": BioGptForCausalLM, """token-classification""": BioGptForTokenClassification, """zero-shot""": BioGptForSequenceClassification, } if is_torch_available() else {} ) __A = False def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = BioGptModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=snake_case__ , hidden_size=37) def __UpperCAmelCase ( self : Any) -> Union[str, Any]: """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : List[str]) -> Dict: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__) def __UpperCAmelCase ( self : Optional[int]) -> int: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*snake_case__) def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case__) def __UpperCAmelCase ( self : Tuple) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case__ , gradient_checkpointing=snake_case__) def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case__) def __UpperCAmelCase ( self : Optional[Any]) -> int: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case__) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case__) @slow def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = BioGptForCausalLM.from_pretrained("microsoft/biogpt") model.to(snake_case__) _UpperCamelCase = BioGptTokenizer.from_pretrained("microsoft/biogpt") _UpperCamelCase = '''left''' # Define PAD Token = EOS Token = 50256 _UpperCamelCase = tokenizer.eos_token _UpperCamelCase = model.config.eos_token_id # use different length sentences to test batching _UpperCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] _UpperCamelCase = tokenizer(snake_case__ , return_tensors="pt" , padding=snake_case__) _UpperCamelCase = inputs['''input_ids'''].to(snake_case__) _UpperCamelCase = model.generate( input_ids=snake_case__ , attention_mask=inputs["attention_mask"].to(snake_case__) , ) _UpperCamelCase = tokenizer(sentences[0] , return_tensors="pt").input_ids.to(snake_case__) _UpperCamelCase = model.generate(input_ids=snake_case__) _UpperCamelCase = inputs_non_padded.shape[-1] - inputs['''attention_mask'''][-1].long().sum().cpu().item() _UpperCamelCase = tokenizer(sentences[1] , return_tensors="pt").input_ids.to(snake_case__) _UpperCamelCase = model.generate(input_ids=snake_case__ , max_length=model.config.max_length - num_paddings) _UpperCamelCase = tokenizer.batch_decode(snake_case__ , skip_special_tokens=snake_case__) _UpperCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case__) _UpperCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case__) _UpperCamelCase = [ '''Hello, my dog is a little bit bigger than a little bit.''', '''Today, I have a good idea of how to use the information''', ] self.assertListEqual(snake_case__ , snake_case__) self.assertListEqual(snake_case__ , [non_padded_sentence, padded_sentence]) @slow def __UpperCAmelCase ( self : Dict) -> int: """simple docstring""" for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = BioGptModel.from_pretrained(snake_case__) self.assertIsNotNone(snake_case__) def __UpperCAmelCase ( self : List[str]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = 3 _UpperCamelCase = input_dict['''input_ids'''] _UpperCamelCase = input_ids.ne(1).to(snake_case__) _UpperCamelCase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) _UpperCamelCase = BioGptForSequenceClassification(snake_case__) model.to(snake_case__) model.eval() _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def __UpperCAmelCase ( self : Optional[Any]) -> int: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = 3 _UpperCamelCase = '''multi_label_classification''' _UpperCamelCase = input_dict['''input_ids'''] _UpperCamelCase = input_ids.ne(1).to(snake_case__) _UpperCamelCase = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) _UpperCamelCase = BioGptForSequenceClassification(snake_case__) model.to(snake_case__) model.eval() _UpperCamelCase = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self : int) -> List[Any]: """simple docstring""" _UpperCamelCase = BioGptForCausalLM.from_pretrained("microsoft/biogpt") _UpperCamelCase = torch.tensor([[2, 4805, 9, 656, 21]]) _UpperCamelCase = model(snake_case__)[0] _UpperCamelCase = 42384 _UpperCamelCase = torch.Size((1, 5, vocab_size)) self.assertEqual(output.shape , snake_case__) _UpperCamelCase = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4)) @slow def __UpperCAmelCase ( self : List[str]) -> Dict: """simple docstring""" _UpperCamelCase = BioGptTokenizer.from_pretrained("microsoft/biogpt") _UpperCamelCase = BioGptForCausalLM.from_pretrained("microsoft/biogpt") model.to(snake_case__) torch.manual_seed(0) _UpperCamelCase = tokenizer("COVID-19 is" , return_tensors="pt").to(snake_case__) _UpperCamelCase = model.generate( **snake_case__ , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case__ , ) _UpperCamelCase = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case__) _UpperCamelCase = ( '''COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the''' ''' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and''' ''' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),''' ''' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and''' ''' more than 800,000 deaths.''' ) self.assertEqual(snake_case__ , snake_case__)
361
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class _UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : Optional[Any] , ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = 13 _UpperCamelCase = 7 _UpperCamelCase = 30 _UpperCamelCase = self.seq_length + self.mem_len _UpperCamelCase = 15 _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = 99 _UpperCamelCase = [10, 50, 80] _UpperCamelCase = 32 _UpperCamelCase = 32 _UpperCamelCase = 4 _UpperCamelCase = 8 _UpperCamelCase = 128 _UpperCamelCase = 2 _UpperCamelCase = 2 _UpperCamelCase = None _UpperCamelCase = 1 _UpperCamelCase = 0 _UpperCamelCase = 3 _UpperCamelCase = self.vocab_size - 1 _UpperCamelCase = 0.01 def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Tuple: """simple docstring""" random.seed(self.seed) tf.random.set_seed(self.seed) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : Optional[Any] , lowercase_ : Optional[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = TFTransfoXLModel(lowercase_) _UpperCamelCase , _UpperCamelCase = model(lowercase_).to_tuple() _UpperCamelCase = {"input_ids": input_ids_a, "mems": mems_a} _UpperCamelCase , _UpperCamelCase = model(lowercase_).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : str , lowercase_ : Dict , lowercase_ : List[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = TFTransfoXLLMHeadModel(lowercase_) _UpperCamelCase , _UpperCamelCase = model(lowercase_).to_tuple() _UpperCamelCase = {"input_ids": input_ids_a, "labels": lm_labels} _UpperCamelCase , _UpperCamelCase = model(lowercase_).to_tuple() _UpperCamelCase , _UpperCamelCase = model([input_ids_a, mems_a]).to_tuple() _UpperCamelCase = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} _UpperCamelCase , _UpperCamelCase = model(lowercase_).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any] , lowercase_ : Dict) -> str: """simple docstring""" _UpperCamelCase = TFTransfoXLForSequenceClassification(lowercase_) _UpperCamelCase = model(lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() ((_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase)) = config_and_inputs _UpperCamelCase = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class _UpperCAmelCase ( lowerCAmelCase, lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) __A = () if is_tf_available() else () __A = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented __A = False __A = False __A = False __A = False def __UpperCAmelCase ( self : List[Any] , lowercase_ : Dict , lowercase_ : Tuple , lowercase_ : Dict , lowercase_ : Any , lowercase_ : List[str]) -> Any: """simple docstring""" if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __UpperCAmelCase ( self : Optional[Any]) -> int: """simple docstring""" _UpperCamelCase = TFTransfoXLModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=lowercase_ , d_embed=37) def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Union[str, Any]) -> List[str]: """simple docstring""" self.model_tester.set_seed() _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*lowercase_) def __UpperCAmelCase ( self : Optional[Any]) -> List[Any]: """simple docstring""" self.model_tester.set_seed() _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*lowercase_) def __UpperCAmelCase ( self : List[str]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*lowercase_) def __UpperCAmelCase ( self : Dict) -> int: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer) if model_class in list_other_models_with_output_ebd: _UpperCamelCase = model.get_output_embeddings() assert isinstance(lowercase_ , tf.keras.layers.Layer) _UpperCamelCase = model.get_bias() assert name is None else: _UpperCamelCase = model.get_output_embeddings() assert x is None _UpperCamelCase = model.get_bias() assert name is None def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" pass @slow def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = TFTransfoXLModel.from_pretrained(lowercase_) self.assertIsNotNone(lowercase_) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss.") def __UpperCAmelCase ( self : Union[str, Any]) -> Tuple: """simple docstring""" pass @require_tf class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @unittest.skip("Skip test until #12651 is resolved.") @slow def __UpperCAmelCase ( self : Optional[Any]) -> Dict: """simple docstring""" _UpperCamelCase = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103") # fmt: off _UpperCamelCase = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off _UpperCamelCase = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> _UpperCamelCase = model.generate(lowercase_ , max_length=200 , do_sample=lowercase_) self.assertListEqual(output_ids[0].numpy().tolist() , lowercase_)
63
0
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" if ( (cp >= 0x4_e_0_0 and cp <= 0x9_f_f_f) or (cp >= 0x3_4_0_0 and cp <= 0x4_d_b_f) # or (cp >= 0x2_0_0_0_0 and cp <= 0x2_a_6_d_f) # or (cp >= 0x2_a_7_0_0 and cp <= 0x2_b_7_3_f) # or (cp >= 0x2_b_7_4_0 and cp <= 0x2_b_8_1_f) # or (cp >= 0x2_b_8_2_0 and cp <= 0x2_c_e_a_f) # or (cp >= 0xf_9_0_0 and cp <= 0xf_a_f_f) or (cp >= 0x2_f_8_0_0 and cp <= 0x2_f_a_1_f) # ): # return True return False def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" for char in word: A__ = ord(_lowerCamelCase ) if not _is_chinese_char(_lowerCamelCase ): return 0 return 1 def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" A__ = set() for token in tokens: A__ = len(_lowerCamelCase ) > 1 and is_chinese(_lowerCamelCase ) if chinese_word: word_set.add(_lowerCamelCase ) A__ = list(_lowerCamelCase ) return word_list def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if not chinese_word_set: return bert_tokens A__ = max([len(_lowerCamelCase ) for w in chinese_word_set] ) A__ = bert_tokens A__ = 0, len(_lowerCamelCase ) while start < end: A__ = True if is_chinese(bert_word[start] ): A__ = min(end - start , _lowerCamelCase ) for i in range(_lowerCamelCase , 1 , -1 ): A__ = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): A__ = "##" + bert_word[j] A__ = start + i A__ = False break if single_word: start += 1 return bert_word def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" A__ = [] for i in range(0 , len(_lowerCamelCase ) , 100 ): A__ = ltp_tokenizer.seg(lines[i : i + 100] )[0] A__ = [get_chinese_word(_lowerCamelCase ) for r in res] ltp_res.extend(_lowerCamelCase ) assert len(_lowerCamelCase ) == len(_lowerCamelCase ) A__ = [] for i in range(0 , len(_lowerCamelCase ) , 100 ): A__ = bert_tokenizer(lines[i : i + 100] , add_special_tokens=_lowerCamelCase , truncation=_lowerCamelCase , max_length=512 ) bert_res.extend(res['input_ids'] ) assert len(_lowerCamelCase ) == len(_lowerCamelCase ) A__ = [] for input_ids, chinese_word in zip(_lowerCamelCase , _lowerCamelCase ): A__ = [] for id in input_ids: A__ = bert_tokenizer._convert_id_to_token(_lowerCamelCase ) input_tokens.append(_lowerCamelCase ) A__ = add_sub_symbol(_lowerCamelCase , _lowerCamelCase ) A__ = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(_lowerCamelCase ): if token[:2] == "##": A__ = token[2:] # save chinese tokens' pos if len(_lowerCamelCase ) == 1 and _is_chinese_char(ord(_lowerCamelCase ) ): ref_id.append(_lowerCamelCase ) ref_ids.append(_lowerCamelCase ) assert len(_lowerCamelCase ) == len(_lowerCamelCase ) return ref_ids def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: A__ = f.readlines() A__ = [line.strip() for line in data if len(_lowerCamelCase ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' A__ = LTP(args.ltp ) # faster in GPU device A__ = BertTokenizer.from_pretrained(args.bert ) A__ = prepare_ref(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: A__ = [json.dumps(_lowerCamelCase ) + "\n" for ref in ref_ids] f.writelines(_lowerCamelCase ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser(description="prepare_chinese_ref") parser.add_argument( "--file_name", type=str, default="./resources/chinese-demo.txt", help="file need process, same as training data in lm", ) parser.add_argument( "--ltp", type=str, default="./resources/ltp", help="resources for LTP tokenizer, usually a path" ) parser.add_argument("--bert", type=str, default="./resources/robert", help="resources for Bert tokenizer") parser.add_argument("--save_path", type=str, default="./resources/ref.txt", help="path to save res") __lowerCamelCase = parser.parse_args() main(args)
221
def A ( _lowerCamelCase ): '''simple docstring''' if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence _lowerCAmelCase : List[str] = gray_code_sequence_string(_lowerCamelCase ) # # convert them to integers for i in range(len(_lowerCamelCase ) ): _lowerCAmelCase : List[str] = int(sequence[i] , 2 ) return sequence def A ( _lowerCamelCase ): '''simple docstring''' if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] _lowerCAmelCase : List[Any] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits _lowerCAmelCase : Optional[int] = gray_code_sequence_string(bit_count - 1 ) _lowerCAmelCase : str = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): _lowerCAmelCase : Dict = "0" + smaller_sequence[i] sequence.append(_lowerCamelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): _lowerCAmelCase : Optional[Any] = "1" + smaller_sequence[i] sequence.append(_lowerCamelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
36
0
"""simple docstring""" import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class __A( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = (DDPMParallelScheduler,) def UpperCAmelCase_ (self , **SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = { "num_train_timesteps": 10_00, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", "variance_type": "fixed_small", "clip_sample": True, } config.update(**_a ) return config def UpperCAmelCase_ (self ): for timesteps in [1, 5, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=_a ) def UpperCAmelCase_ (self ): for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=_a , beta_end=_a ) def UpperCAmelCase_ (self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_a ) def UpperCAmelCase_ (self ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=_a ) def UpperCAmelCase_ (self ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=_a ) def UpperCAmelCase_ (self ): self.check_over_configs(thresholding=_a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=_a , prediction_type=_a , sample_max_value=_a , ) def UpperCAmelCase_ (self ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=_a ) def UpperCAmelCase_ (self ): for t in [0, 5_00, 9_99]: self.check_over_forward(time_step=_a ) def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 ) - 0.02 ) ) < 1E-5 def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = len(_a ) UpperCamelCase__ = self.dummy_model() UpperCamelCase__ = self.dummy_sample_deter UpperCamelCase__ = self.dummy_sample_deter + 0.1 UpperCamelCase__ = self.dummy_sample_deter - 0.1 UpperCamelCase__ = samplea.shape[0] UpperCamelCase__ = torch.stack([samplea, samplea, samplea] , dim=0 ) UpperCamelCase__ = torch.arange(_a )[0:3, None].repeat(1 , _a ) UpperCamelCase__ = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) UpperCamelCase__ = scheduler.batch_step_no_noise(_a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) UpperCamelCase__ = torch.sum(torch.abs(_a ) ) UpperCamelCase__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 11_53.18_33 ) < 1E-2 assert abs(result_mean.item() - 0.5005 ) < 1E-3 def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = len(_a ) UpperCamelCase__ = self.dummy_model() UpperCamelCase__ = self.dummy_sample_deter UpperCamelCase__ = torch.manual_seed(0 ) for t in reversed(range(_a ) ): # 1. predict noise residual UpperCamelCase__ = model(_a , _a ) # 2. predict previous mean of sample x_t-1 UpperCamelCase__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample UpperCamelCase__ = pred_prev_sample UpperCamelCase__ = torch.sum(torch.abs(_a ) ) UpperCamelCase__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 2_58.96_06 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config(prediction_type="""v_prediction""" ) UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = len(_a ) UpperCamelCase__ = self.dummy_model() UpperCamelCase__ = self.dummy_sample_deter UpperCamelCase__ = torch.manual_seed(0 ) for t in reversed(range(_a ) ): # 1. predict noise residual UpperCamelCase__ = model(_a , _a ) # 2. predict previous mean of sample x_t-1 UpperCamelCase__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample UpperCamelCase__ = pred_prev_sample UpperCamelCase__ = torch.sum(torch.abs(_a ) ) UpperCamelCase__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 2_02.02_96 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = [1_00, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=_a ) UpperCamelCase__ = scheduler.timesteps for i, timestep in enumerate(_a ): if i == len(_a ) - 1: UpperCamelCase__ = -1 else: UpperCamelCase__ = timesteps[i + 1] UpperCamelCase__ = scheduler.previous_timestep(_a ) UpperCamelCase__ = prev_t.item() self.assertEqual(_a , _a ) def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = [1_00, 87, 50, 51, 0] with self.assertRaises(_a , msg="""`custom_timesteps` must be in descending order.""" ): scheduler.set_timesteps(timesteps=_a ) def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = [1_00, 87, 50, 1, 0] UpperCamelCase__ = len(_a ) with self.assertRaises(_a , msg="""Can only pass one of `num_inference_steps` or `custom_timesteps`.""" ): scheduler.set_timesteps(num_inference_steps=_a , timesteps=_a ) def UpperCAmelCase_ (self ): UpperCamelCase__ = self.scheduler_classes[0] UpperCamelCase__ = self.get_scheduler_config() UpperCamelCase__ = scheduler_class(**_a ) UpperCamelCase__ = [scheduler.config.num_train_timesteps] with self.assertRaises( _a , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=_a )
369
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase_ = { '''configuration_canine''': ['''CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CanineConfig'''], '''tokenization_canine''': ['''CanineTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''CANINE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CanineForMultipleChoice''', '''CanineForQuestionAnswering''', '''CanineForSequenceClassification''', '''CanineForTokenClassification''', '''CanineLayer''', '''CanineModel''', '''CaninePreTrainedModel''', '''load_tf_weights_in_canine''', ] if TYPE_CHECKING: from .configuration_canine import CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP, CanineConfig from .tokenization_canine import CanineTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_canine import ( CANINE_PRETRAINED_MODEL_ARCHIVE_LIST, CanineForMultipleChoice, CanineForQuestionAnswering, CanineForSequenceClassification, CanineForTokenClassification, CanineLayer, CanineModel, CaninePreTrainedModel, load_tf_weights_in_canine, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
178
0
"""simple docstring""" import math def UpperCamelCase_ ( lowerCAmelCase__ : float , lowerCAmelCase__ : float ) -> float: """simple docstring""" return math.pow(a__ , 2 ) - a def UpperCamelCase_ ( lowerCAmelCase__ : float ) -> float: """simple docstring""" return 2 * x def UpperCamelCase_ ( lowerCAmelCase__ : float ) -> float: """simple docstring""" lowerCAmelCase_ : Tuple = 2.0 while start <= a: lowerCAmelCase_ : Dict = math.pow(a__ , 2 ) return start def UpperCamelCase_ ( lowerCAmelCase__ : float , lowerCAmelCase__ : int = 9999 , lowerCAmelCase__ : float = 0.00000000000001 ) -> float: """simple docstring""" if a < 0: raise ValueError('math domain error' ) lowerCAmelCase_ : str = get_initial_point(a__ ) for _ in range(a__ ): lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = value - fx(a__ , a__ ) / fx_derivative(a__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
224
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCAmelCase = logging.get_logger(__name__) class UpperCAmelCase_ ( _lowercase): snake_case__ = ['''input_values''', '''padding_mask'''] def __init__( self : Optional[Any] , __UpperCamelCase : int = 1 , __UpperCamelCase : int = 2_4000 , __UpperCamelCase : float = 0.0 , __UpperCamelCase : float = None , __UpperCamelCase : float = None , **__UpperCamelCase : Optional[Any] , ) -> Optional[int]: super().__init__(feature_size=__UpperCamelCase , sampling_rate=__UpperCamelCase , padding_value=__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = chunk_length_s _UpperCamelCase = overlap @property def _UpperCamelCase ( self : Optional[int] ) -> Optional[int]: if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _UpperCamelCase ( self : Union[str, Any] ) -> Optional[int]: if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self : Union[str, Any] , __UpperCamelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCamelCase : Optional[Union[bool, str, PaddingStrategy]] = None , __UpperCamelCase : Optional[bool] = False , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : Optional[Union[str, TensorType]] = None , __UpperCamelCase : Optional[int] = None , ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) if padding and truncation: raise ValueError('''Both padding and truncation were set. Make sure you only set one.''' ) elif padding is None: # by default let's pad the inputs _UpperCamelCase = True _UpperCamelCase = bool( isinstance(__UpperCamelCase , (list, tuple) ) and (isinstance(raw_audio[0] , (np.ndarray, tuple, list) )) ) if is_batched: _UpperCamelCase = [np.asarray(__UpperCamelCase , dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(__UpperCamelCase , np.ndarray ): _UpperCamelCase = np.asarray(__UpperCamelCase , dtype=np.floataa ) elif isinstance(__UpperCamelCase , np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): _UpperCamelCase = raw_audio.astype(np.floataa ) # always return batch if not is_batched: _UpperCamelCase = [np.asarray(__UpperCamelCase ).T] # verify inputs are valid for idx, example in enumerate(__UpperCamelCase ): if example.ndim > 2: raise ValueError(F'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(F'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(F'''Expected stereo audio but example has {example.shape[-1]} channels''' ) _UpperCamelCase = None _UpperCamelCase = BatchFeature({'''input_values''': raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: _UpperCamelCase = min(array.shape[0] for array in raw_audio ) _UpperCamelCase = int(np.floor(max_length / self.chunk_stride ) ) _UpperCamelCase = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: _UpperCamelCase = max(array.shape[0] for array in raw_audio ) _UpperCamelCase = int(np.ceil(max_length / self.chunk_stride ) ) _UpperCamelCase = (nb_step - 1) * self.chunk_stride + self.chunk_length _UpperCamelCase = '''max_length''' else: _UpperCamelCase = input_values # normal padding on batch if padded_inputs is None: _UpperCamelCase = self.pad( __UpperCamelCase , max_length=__UpperCamelCase , truncation=__UpperCamelCase , padding=__UpperCamelCase , return_attention_mask=__UpperCamelCase , ) if padding: _UpperCamelCase = padded_inputs.pop('''attention_mask''' ) _UpperCamelCase = [] for example in padded_inputs.pop('''input_values''' ): if self.feature_size == 1: _UpperCamelCase = example[..., None] input_values.append(example.T ) _UpperCamelCase = input_values if return_tensors is not None: _UpperCamelCase = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs
256
0
'''simple docstring''' import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case__ : Tuple = TapasConfig.from_json_file(__lowerCamelCase ) # set absolute/relative position embeddings parameter snake_case__ : Union[str, Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case__ : str = TapasForQuestionAnswering(config=__lowerCamelCase ) elif task == "WTQ": # run_task_main.py hparams snake_case__ : List[Any] = 4 snake_case__ : Dict = True # hparam_utils.py hparams snake_case__ : Optional[int] = 0.66_4694 snake_case__ : List[str] = 0.20_7951 snake_case__ : Union[str, Any] = 0.12_1194 snake_case__ : Dict = True snake_case__ : Union[str, Any] = True snake_case__ : List[Any] = False snake_case__ : str = 0.035_2513 snake_case__ : List[Any] = TapasForQuestionAnswering(config=__lowerCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case__ : Dict = 4 snake_case__ : Union[str, Any] = False # hparam_utils.py hparams snake_case__ : Optional[int] = 36.4519 snake_case__ : List[Any] = 0.90_3421 snake_case__ : Union[str, Any] = 222.088 snake_case__ : Tuple = True snake_case__ : Any = True snake_case__ : Dict = True snake_case__ : str = 0.76_3141 snake_case__ : Dict = TapasForQuestionAnswering(config=__lowerCamelCase ) elif task == "TABFACT": snake_case__ : List[str] = TapasForSequenceClassification(config=__lowerCamelCase ) elif task == "MLM": snake_case__ : Optional[int] = TapasForMaskedLM(config=__lowerCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": snake_case__ : Any = TapasModel(config=__lowerCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(__lowerCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) snake_case__ : Dict = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + """vocab.txt""" , model_max_length=512 ) tokenizer.save_pretrained(__lowerCamelCase ) print("""Used relative position embeddings:""" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "--task", default="SQA", type=str, help="Model task for which to convert a checkpoint. Defaults to SQA." ) parser.add_argument( "--reset_position_index_per_cell", default=False, action="store_true", help="Whether to use relative position embeddings or not. Defaults to True.", ) parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--tapas_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained TAPAS model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __a = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
351
'''simple docstring''' import re from filelock import FileLock try: import nltk __a = True except (ImportError, ModuleNotFoundError): __a = False if NLTK_AVAILABLE: with FileLock(".lock") as lock: nltk.download("punkt", quiet=True) def __snake_case( _lowerCAmelCase ) -> str: re.sub("""<n>""" , """""" , _lowerCAmelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(_lowerCAmelCase ) )
43
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class _A ( unittest.TestCase ): def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100]) __a = get_activation('''gelu''') self.assertTrue(torch.allclose(gelu_python(__UpperCamelCase) , torch_builtin(__UpperCamelCase))) self.assertFalse(torch.allclose(gelu_python(__UpperCamelCase) , gelu_new(__UpperCamelCase))) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __a = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100]) __a = get_activation('''gelu''') __a = get_activation('''gelu_10''') __a = torch_builtin(__UpperCamelCase) __a = geluaa(__UpperCamelCase) __a = torch.where(y_gelu_aa < 10.0 , 1 , 0) self.assertTrue(torch.max(__UpperCamelCase).item() == 10.0) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask)) def _lowerCamelCase ( self : Dict): '''simple docstring''' get_activation('''gelu''') get_activation('''gelu_10''') get_activation('''gelu_fast''') get_activation('''gelu_new''') get_activation('''gelu_python''') get_activation('''gelu_pytorch_tanh''') get_activation('''linear''') get_activation('''mish''') get_activation('''quick_gelu''') get_activation('''relu''') get_activation('''sigmoid''') get_activation('''silu''') get_activation('''swish''') get_activation('''tanh''') with self.assertRaises(__UpperCamelCase): get_activation('''bogus''') with self.assertRaises(__UpperCamelCase): get_activation(__UpperCamelCase) def _lowerCamelCase ( self : str): '''simple docstring''' __a = get_activation('''gelu''') __a = 1 __a = get_activation('''gelu''') self.assertEqual(acta.a , 1) with self.assertRaises(__UpperCamelCase): __a = acta.a
49
"""simple docstring""" from __future__ import annotations _snake_case : str = [] def A__ ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): for i in range(len(UpperCamelCase ) ): if board[row][i] == 1: return False for i in range(len(UpperCamelCase ) ): if board[i][column] == 1: return False for i, j in zip(range(UpperCamelCase , -1 , -1 ) , range(UpperCamelCase , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(UpperCamelCase , -1 , -1 ) , range(UpperCamelCase , len(UpperCamelCase ) ) ): if board[i][j] == 1: return False return True def A__ ( UpperCamelCase , UpperCamelCase ): if row >= len(UpperCamelCase ): solution.append(UpperCamelCase ) printboard(UpperCamelCase ) print() return True for i in range(len(UpperCamelCase ) ): if is_safe(UpperCamelCase , UpperCamelCase , UpperCamelCase ): A = 1 solve(UpperCamelCase , row + 1 ) A = 0 return False def A__ ( UpperCamelCase ): for i in range(len(UpperCamelCase ) ): for j in range(len(UpperCamelCase ) ): if board[i][j] == 1: print("Q" , end=" " ) else: print("." , end=" " ) print() # n=int(input("The no. of queens")) _snake_case : List[str] = 8 _snake_case : List[str] = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
292
0
from manim import * class snake_case_ ( __lowercase ): def UpperCAmelCase__ ( self : List[Any] )->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = Rectangle(height=0.5 , width=0.5 ) __lowerCAmelCase : Dict = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __lowerCAmelCase : Optional[int] = Rectangle(height=0.25 , width=0.25 ) __lowerCAmelCase : Dict = [mem.copy() for i in range(6 )] __lowerCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] __lowerCAmelCase : Any = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Tuple = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Union[str, Any] = VGroup(_snake_case , _snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Optional[int] = Text("""CPU""" , font_size=24 ) __lowerCAmelCase : List[str] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_snake_case ) __lowerCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] __lowerCAmelCase : List[str] = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : str = Text("""GPU""" , font_size=24 ) __lowerCAmelCase : Optional[int] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) gpu.move_to([-1, -1, 0] ) self.add(_snake_case ) __lowerCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] __lowerCAmelCase : Any = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Union[str, Any] = Text("""Model""" , font_size=24 ) __lowerCAmelCase : Optional[Any] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) model.move_to([3, -1.0, 0] ) self.add(_snake_case ) __lowerCAmelCase : Dict = [] __lowerCAmelCase : str = [] for i, rect in enumerate(_snake_case ): __lowerCAmelCase : Optional[int] = fill.copy().set_fill(_snake_case , opacity=0.8 ) target.move_to(_snake_case ) model_arr.append(_snake_case ) __lowerCAmelCase : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(_snake_case , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(_snake_case ) self.add(*_snake_case , *_snake_case ) __lowerCAmelCase : str = [meta_mem.copy() for i in range(6 )] __lowerCAmelCase : Tuple = [meta_mem.copy() for i in range(6 )] __lowerCAmelCase : str = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Dict = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : str = VGroup(_snake_case , _snake_case ).arrange(_snake_case , buff=0 ) __lowerCAmelCase : Optional[Any] = Text("""Disk""" , font_size=24 ) __lowerCAmelCase : str = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) disk.move_to([-4, -1.25, 0] ) self.add(_snake_case , _snake_case ) __lowerCAmelCase : Tuple = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __lowerCAmelCase : str = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_snake_case , _snake_case ) __lowerCAmelCase : Optional[int] = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_snake_case ) __lowerCAmelCase : Dict = MarkupText( F'''Now watch as an input is passed through the model\nand how the memory is utilized and handled.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case ) ) __lowerCAmelCase : List[Any] = Square(0.3 ) input.set_fill(_snake_case , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , _snake_case , buff=0.5 ) self.play(Write(_snake_case ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=_snake_case , buff=0.02 ) self.play(MoveToTarget(_snake_case ) ) self.play(FadeOut(_snake_case ) ) __lowerCAmelCase : int = Arrow(start=_snake_case , end=_snake_case , color=_snake_case , buff=0.5 ) a.next_to(model_arr[0].get_left() , _snake_case , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) __lowerCAmelCase : Optional[Any] = MarkupText( F'''As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case , run_time=3 ) ) __lowerCAmelCase : int = {"""run_time""": 1, """fade_in""": True, """fade_out""": True, """buff""": 0.02} self.play( Write(_snake_case ) , Circumscribe(model_arr[0] , color=_snake_case , **_snake_case ) , Circumscribe(model_cpu_arr[0] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) __lowerCAmelCase : Optional[Any] = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , _snake_case , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) __lowerCAmelCase : str = AnimationGroup( FadeOut(_snake_case , run_time=0.5 ) , MoveToTarget(_snake_case , run_time=0.5 ) , FadeIn(_snake_case , run_time=0.5 ) , lag_ratio=0.2 ) self.play(_snake_case ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: __lowerCAmelCase : Optional[int] = 0.7 self.play( Circumscribe(model_arr[i] , **_snake_case ) , Circumscribe(cpu_left_col_base[i] , **_snake_case ) , Circumscribe(cpu_left_col_base[i + 1] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , Circumscribe(model_arr[i + 1] , color=_snake_case , **_snake_case ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=_snake_case , **_snake_case ) , Circumscribe(cpu_left_col_base[-1] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) __lowerCAmelCase : List[Any] = a_c __lowerCAmelCase : Optional[Any] = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(_snake_case ) , FadeOut(_snake_case , run_time=0.5 ) , ) __lowerCAmelCase : Dict = MarkupText(F'''Inference on a model too large for GPU memory\nis successfully completed.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case , run_time=3 ) , MoveToTarget(_snake_case ) ) self.wait()
232
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case_ ( __lowercase ): def UpperCAmelCase__ ( self : Dict )->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_snake_case , """embed_dim""" ) ) self.parent.assertTrue(hasattr(_snake_case , """num_heads""" ) ) class snake_case_ : def __init__( self : Dict , _snake_case : int , _snake_case : str=13 , _snake_case : Optional[int]=64 , _snake_case : Union[str, Any]=3 , _snake_case : Any=[16, 48, 96] , _snake_case : List[str]=[1, 3, 6] , _snake_case : str=[1, 2, 10] , _snake_case : Tuple=[7, 3, 3] , _snake_case : Tuple=[4, 2, 2] , _snake_case : Tuple=[2, 1, 1] , _snake_case : List[str]=[2, 2, 2] , _snake_case : Tuple=[False, False, True] , _snake_case : int=[0.0, 0.0, 0.0] , _snake_case : Union[str, Any]=0.02 , _snake_case : List[str]=1E-12 , _snake_case : str=True , _snake_case : Any=True , _snake_case : Optional[Any]=2 , )->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = parent __lowerCAmelCase : int = batch_size __lowerCAmelCase : Optional[int] = image_size __lowerCAmelCase : Optional[Any] = patch_sizes __lowerCAmelCase : Tuple = patch_stride __lowerCAmelCase : List[Any] = patch_padding __lowerCAmelCase : Tuple = is_training __lowerCAmelCase : str = use_labels __lowerCAmelCase : List[Any] = num_labels __lowerCAmelCase : int = num_channels __lowerCAmelCase : Tuple = embed_dim __lowerCAmelCase : Optional[int] = num_heads __lowerCAmelCase : Union[str, Any] = stride_kv __lowerCAmelCase : List[Any] = depth __lowerCAmelCase : int = cls_token __lowerCAmelCase : Optional[Any] = attention_drop_rate __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Any = layer_norm_eps def UpperCAmelCase__ ( self : List[str] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase : Optional[int] = None if self.use_labels: # create a random int32 tensor of given shape __lowerCAmelCase : Tuple = ids_tensor([self.batch_size] , self.num_labels ) __lowerCAmelCase : List[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase__ ( self : List[str] )->int: '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self : List[Any] , _snake_case : int , _snake_case : str , _snake_case : Union[str, Any] )->Tuple: '''simple docstring''' __lowerCAmelCase : str = TFCvtModel(config=_snake_case ) __lowerCAmelCase : Optional[Any] = model(_snake_case , training=_snake_case ) __lowerCAmelCase : str = (self.image_size, self.image_size) __lowerCAmelCase , __lowerCAmelCase : Tuple = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowerCAmelCase : int = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowerCAmelCase : Any = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def UpperCAmelCase__ ( self : Tuple , _snake_case : Optional[int] , _snake_case : Tuple , _snake_case : Optional[Any] )->Dict: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.num_labels __lowerCAmelCase : Optional[int] = TFCvtForImageClassification(_snake_case ) __lowerCAmelCase : str = model(_snake_case , labels=_snake_case , training=_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Tuple )->str: '''simple docstring''' __lowerCAmelCase : Tuple = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Any = config_and_inputs __lowerCAmelCase : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class snake_case_ ( __lowercase ,__lowercase ,unittest.TestCase ): A_ = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () A_ = ( {'feature-extraction': TFCvtModel, 'image-classification': TFCvtForImageClassification} if is_tf_available() else {} ) A_ = False A_ = False A_ = False A_ = False A_ = False def UpperCAmelCase__ ( self : List[str] )->str: '''simple docstring''' __lowerCAmelCase : Tuple = TFCvtModelTester(self ) __lowerCAmelCase : Optional[Any] = TFCvtConfigTester(self , config_class=_snake_case , has_text_modality=_snake_case , hidden_size=37 ) def UpperCAmelCase__ ( self : Union[str, Any] )->Optional[int]: '''simple docstring''' self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="""Cvt does not output attentions""" ) def UpperCAmelCase__ ( self : str )->List[Any]: '''simple docstring''' pass @unittest.skip(reason="""Cvt does not use inputs_embeds""" ) def UpperCAmelCase__ ( self : Union[str, Any] )->List[str]: '''simple docstring''' pass @unittest.skip(reason="""Cvt does not support input and output embeddings""" ) def UpperCAmelCase__ ( self : Tuple )->Optional[int]: '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) def UpperCAmelCase__ ( self : Dict )->Any: '''simple docstring''' super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase__ ( self : Dict )->Dict: '''simple docstring''' super().test_keras_fit() @unittest.skip(reason="""Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8""" ) def UpperCAmelCase__ ( self : Union[str, Any] )->str: '''simple docstring''' __lowerCAmelCase : Optional[int] = tf.keras.mixed_precision.Policy("""mixed_float16""" ) tf.keras.mixed_precision.set_global_policy(_snake_case ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("""float32""" ) def UpperCAmelCase__ ( self : Tuple )->Tuple: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : Union[str, Any] = model_class(_snake_case ) __lowerCAmelCase : Tuple = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase : int = [*signature.parameters.keys()] __lowerCAmelCase : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _snake_case ) def UpperCAmelCase__ ( self : int )->List[str]: '''simple docstring''' def check_hidden_states_output(_snake_case : Union[str, Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ): __lowerCAmelCase : Any = model_class(_snake_case ) __lowerCAmelCase : Any = model(**self._prepare_for_class(_snake_case , _snake_case ) ) __lowerCAmelCase : Optional[Any] = outputs.hidden_states __lowerCAmelCase : Tuple = len(self.model_tester.depth ) self.assertEqual(len(_snake_case ) , _snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : str = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCAmelCase : Optional[Any] = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) def UpperCAmelCase__ ( self : str )->List[str]: '''simple docstring''' __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCAmelCase__ ( self : Dict )->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_snake_case ) @slow def UpperCAmelCase__ ( self : Dict )->Union[str, Any]: '''simple docstring''' for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : List[Any] = TFCvtModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def _SCREAMING_SNAKE_CASE ( ) -> Tuple: __lowerCAmelCase : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class snake_case_ ( unittest.TestCase ): @cached_property def UpperCAmelCase__ ( self : Dict )->List[Any]: '''simple docstring''' return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def UpperCAmelCase__ ( self : List[str] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Any = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __lowerCAmelCase : List[Any] = self.default_image_processor __lowerCAmelCase : Optional[int] = prepare_img() __lowerCAmelCase : int = image_processor(images=_snake_case , return_tensors="""tf""" ) # forward pass __lowerCAmelCase : Dict = model(**_snake_case ) # verify the logits __lowerCAmelCase : Dict = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _snake_case ) __lowerCAmelCase : Any = tf.constant([0.9_285, 0.9_015, -0.3_150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _snake_case , atol=1E-4 ) )
232
1
from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : str = { '''nielsr/canine-s''': 2048, } # Unicode defines 1,114,112 total “codepoints” _SCREAMING_SNAKE_CASE : Any = 1114112 # Below: Constants defining canonical codepoints for special, pseudo-characters. # Copied from https://github.com/google-research/language/blob/master/language/canine/special_codepoints.py _SCREAMING_SNAKE_CASE : List[Any] = 0 _SCREAMING_SNAKE_CASE : Tuple = 0xE_0_0_0 _SCREAMING_SNAKE_CASE : str = 0xE_0_0_1 _SCREAMING_SNAKE_CASE : List[str] = 0xE_0_0_2 _SCREAMING_SNAKE_CASE : Union[str, Any] = 0xE_0_0_3 _SCREAMING_SNAKE_CASE : Optional[Any] = 0xE_0_0_4 # Maps special codepoints to human-readable names. _SCREAMING_SNAKE_CASE : Dict[int, str] = { # Special symbols are represented using codepoints values that are valid, # but designated as "Private Use", meaning that they will never be assigned # characters by the Unicode Consortium, and are thus safe for use here. # # NOTE: Do *NOT* add any sort of [UNK_CHAR] here. They are explicitly # excluded and should fail with a hard error. CLS: "[CLS]", SEP: "[SEP]", BOS: "[BOS]", MASK: "[MASK]", PAD: "[PAD]", RESERVED: "[RESERVED]", } # Maps special codepoint human-readable names to their codepoint values. _SCREAMING_SNAKE_CASE : Dict[str, int] = {name: codepoint for codepoint, name in SPECIAL_CODEPOINTS.items()} class UpperCAmelCase__ ( A__ ): """simple docstring""" a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , __lowerCamelCase : List[Any]=chr(__lowerCamelCase ) , __lowerCamelCase : List[Any]=chr(__lowerCamelCase ) , __lowerCamelCase : List[str]=chr(__lowerCamelCase ) , __lowerCamelCase : List[Any]=chr(__lowerCamelCase ) , __lowerCamelCase : int=chr(__lowerCamelCase ) , __lowerCamelCase : str=chr(__lowerCamelCase ) , __lowerCamelCase : int=False , __lowerCamelCase : int=2048 , **__lowerCamelCase : Optional[int] , ) -> int: SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else bos_token SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else eos_token SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else sep_token SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else cls_token SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE__ = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else mask_token super().__init__( bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , sep_token=__lowerCamelCase , cls_token=__lowerCamelCase , pad_token=__lowerCamelCase , mask_token=__lowerCamelCase , add_prefix_space=__lowerCamelCase , model_max_length=__lowerCamelCase , **__lowerCamelCase , ) # Creates a mapping for looking up the IDs of special symbols. SCREAMING_SNAKE_CASE__ = {} for codepoint, name in SPECIAL_CODEPOINTS.items(): SCREAMING_SNAKE_CASE__ = codepoint # Creates a mapping for looking up the string forms of special symbol IDs. SCREAMING_SNAKE_CASE__ = { codepoint: name for name, codepoint in self._special_codepoints.items() } SCREAMING_SNAKE_CASE__ = UNICODE_VOCAB_SIZE SCREAMING_SNAKE_CASE__ = len(self._special_codepoints ) @property def lowercase_ ( self : Tuple ) -> int: return self._unicode_vocab_size def lowercase_ ( self : Dict , __lowerCamelCase : str ) -> List[str]: return list(__lowerCamelCase ) def lowercase_ ( self : Optional[Any] , __lowerCamelCase : str ) -> int: try: return ord(__lowerCamelCase ) except TypeError: raise ValueError(f'''invalid token: \'{token}\'''' ) def lowercase_ ( self : int , __lowerCamelCase : int ) -> str: try: if index in SPECIAL_CODEPOINTS: return SPECIAL_CODEPOINTS[index] return chr(__lowerCamelCase ) except TypeError: raise ValueError(f'''invalid id: {index}''' ) def lowercase_ ( self : Tuple , __lowerCamelCase : Tuple ) -> List[str]: return "".join(__lowerCamelCase ) def lowercase_ ( self : List[str] , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ) -> List[int]: SCREAMING_SNAKE_CASE__ = [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] SCREAMING_SNAKE_CASE__ = cls + token_ids_a + sep if token_ids_a is not None: result += token_ids_a + sep return result def lowercase_ ( self : Tuple , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [1] + ([0] * len(__lowerCamelCase )) + [1] if token_ids_a is not None: result += ([0] * len(__lowerCamelCase )) + [1] return result def lowercase_ ( self : Union[str, Any] , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ) -> List[int]: SCREAMING_SNAKE_CASE__ = [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] SCREAMING_SNAKE_CASE__ = len(cls + token_ids_a + sep ) * [0] if token_ids_a is not None: result += len(token_ids_a + sep ) * [1] return result def lowercase_ ( self : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ) -> Tuple: return ()
314
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self : List[str] ) -> Tuple: SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) SCREAMING_SNAKE_CASE__ = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48145466, 0.4578275, 0.40821073], '''image_std''': [0.26862954, 0.26130258, 0.27577711], } SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def lowercase_ ( self : List[str] , **__lowerCamelCase : Dict ) -> List[str]: return BertTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowercase_ ( self : Any , **__lowerCamelCase : List[str] ) -> Any: return BertTokenizerFast.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowercase_ ( self : Optional[int] , **__lowerCamelCase : int ) -> Dict: return EfficientNetImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowercase_ ( self : Dict ) -> Dict: shutil.rmtree(self.tmpdirname ) def lowercase_ ( self : List[Any] ) -> Dict: SCREAMING_SNAKE_CASE__ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] SCREAMING_SNAKE_CASE__ = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowercase_ ( self : int ) -> str: SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = AlignProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def lowercase_ ( self : Optional[int] ) -> List[str]: SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) SCREAMING_SNAKE_CASE__ = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) SCREAMING_SNAKE_CASE__ = AlignProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def lowercase_ ( self : Optional[Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = image_processor(__lowerCamelCase , return_tensors='''np''' ) SCREAMING_SNAKE_CASE__ = processor(images=__lowerCamelCase , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowercase_ ( self : Tuple ) -> List[Any]: SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = '''lower newer''' SCREAMING_SNAKE_CASE__ = processor(text=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = tokenizer(__lowerCamelCase , padding='''max_length''' , max_length=64 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowercase_ ( self : Optional[int] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = '''lower newer''' SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def lowercase_ ( self : Union[str, Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE__ = processor.batch_decode(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def lowercase_ ( self : int ) -> str: SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = AlignProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = '''lower newer''' SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
314
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { '''configuration_mask2former''': [ '''MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Mask2FormerConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ['''Mask2FormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Mask2FormerForUniversalSegmentation''', '''Mask2FormerModel''', '''Mask2FormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
334
'''simple docstring''' import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowerCAmelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase_ : Optional[Any] = DebertaTokenizer UpperCamelCase_ : List[str] = True UpperCamelCase_ : int = DebertaTokenizerFast def _snake_case ( self : Optional[int] ) -> Dict: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A: Optional[int] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''[UNK]''', ] A: int = dict(zip(SCREAMING_SNAKE_CASE_ , range(len(SCREAMING_SNAKE_CASE_ ) ) ) ) A: Union[str, Any] = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] A: Union[str, Any] = {'''unk_token''': '''[UNK]'''} A: List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) A: str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(SCREAMING_SNAKE_CASE_ ) ) def _snake_case ( self : int , **SCREAMING_SNAKE_CASE_ : List[str] ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : Dict ) -> Union[str, Any]: '''simple docstring''' A: Optional[int] = '''lower newer''' A: str = '''lower newer''' return input_text, output_text def _snake_case ( self : Union[str, Any] ) -> Dict: '''simple docstring''' A: str = self.get_tokenizer() A: Any = '''lower newer''' A: Dict = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] A: int = tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) A: List[Any] = tokens + [tokenizer.unk_token] A: int = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : List[Any] ) -> Any: '''simple docstring''' A: str = self.get_tokenizer() A: List[str] = tokenizer('''Hello''' , '''World''' ) A: Union[str, Any] = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd['''token_type_ids'''] , SCREAMING_SNAKE_CASE_ ) @slow def _snake_case ( self : Tuple ) -> Optional[int]: '''simple docstring''' A: Union[str, Any] = self.tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A: Any = tokenizer.encode('''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) A: Union[str, Any] = tokenizer.encode('''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE_ ) A: Dict = tokenizer.encode( '''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE_ , add_prefix_space=SCREAMING_SNAKE_CASE_ ) A: Dict = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE_ , add_prefix_space=SCREAMING_SNAKE_CASE_ ) A: List[Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ ) A: int = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def _snake_case ( self : Tuple ) -> Dict: '''simple docstring''' A: int = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: A: List[Any] = tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A: Dict = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] A: Dict = tokenizer(SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ ) A: Any = [tokenizer.decode(SCREAMING_SNAKE_CASE_ , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) for seq in encoding['''input_ids''']] # fmt: off A: Any = { '''input_ids''': [ [1, 21_18, 1_11_26, 5_65, 35, 83, 2_51_91, 1_63, 1_88_54, 13, 1_21_56, 12, 1_61_01, 2_53_76, 1_38_07, 9, 2_22_05, 2_78_93, 16_35, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 21_18, 1_11_26, 5_65, 2_45_36, 80, 4_37_97, 48_78, 73_73, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1_33, 78, 65, 16, 10, 37_24, 15_38, 3_31_83, 1_13_03, 4_37_97, 19_38, 4, 8_70, 2_41_65, 2_91_05, 5, 7_39, 3_26_44, 3_31_83, 1_13_03, 3_61_73, 88, 80, 6_50, 78_21, 4_59_40, 6, 52, 25_59, 5, 18_36, 9, 5, 73_97, 1_31_71, 31, 5, 18_36, 9, 3_26_44, 3_31_83, 1_13_03, 4, 2] ], '''token_type_ids''': [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], '''attention_mask''': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on A: Optional[int] = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] self.assertDictEqual(encoding.data , SCREAMING_SNAKE_CASE_ ) for expected, decoded in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
334
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = { "configuration_electra": ["ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP", "ElectraConfig", "ElectraOnnxConfig"], "tokenization_electra": ["ElectraTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["ElectraTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "ElectraForCausalLM", "ElectraForMaskedLM", "ElectraForMultipleChoice", "ElectraForPreTraining", "ElectraForQuestionAnswering", "ElectraForSequenceClassification", "ElectraForTokenClassification", "ElectraModel", "ElectraPreTrainedModel", "load_tf_weights_in_electra", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFElectraForMaskedLM", "TFElectraForMultipleChoice", "TFElectraForPreTraining", "TFElectraForQuestionAnswering", "TFElectraForSequenceClassification", "TFElectraForTokenClassification", "TFElectraModel", "TFElectraPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "FlaxElectraForCausalLM", "FlaxElectraForMaskedLM", "FlaxElectraForMultipleChoice", "FlaxElectraForPreTraining", "FlaxElectraForQuestionAnswering", "FlaxElectraForSequenceClassification", "FlaxElectraForTokenClassification", "FlaxElectraModel", "FlaxElectraPreTrainedModel", ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
35
'''simple docstring''' import numpy as np from transformers import Pipeline def __snake_case( _lowerCAmelCase ) -> Optional[int]: snake_case__ : Optional[Any] = np.max(_lowerCAmelCase , axis=-1 , keepdims=_lowerCAmelCase ) snake_case__ : List[str] = np.exp(outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=_lowerCAmelCase ) class UpperCAmelCase_ ( _a ): """simple docstring""" def lowerCamelCase ( self : Optional[Any] , **snake_case_ : int ): snake_case__ : Optional[int] = {} if "second_text" in kwargs: snake_case__ : Union[str, Any] = kwargs["""second_text"""] return preprocess_kwargs, {}, {} def lowerCamelCase ( self : str , snake_case_ : Tuple , snake_case_ : Union[str, Any]=None ): return self.tokenizer(snake_case_ , text_pair=snake_case_ , return_tensors=self.framework ) def lowerCamelCase ( self : List[Any] , snake_case_ : Dict ): return self.model(**snake_case_ ) def lowerCamelCase ( self : int , snake_case_ : List[Any] ): snake_case__ : Union[str, Any] = model_outputs.logits[0].numpy() snake_case__ : List[str] = softmax(snake_case_ ) snake_case__ : List[str] = np.argmax(snake_case_ ) snake_case__ : List[str] = self.model.config.idalabel[best_class] snake_case__ : Optional[int] = probabilities[best_class].item() snake_case__ : str = logits.tolist() return {"label": label, "score": score, "logits": logits}
35
1
'''simple docstring''' from __future__ import annotations def lowerCamelCase__ ( _A , _A , _A , ): if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('You cannot supply more or less than 2 values' ) elif electron_conc < 0: raise ValueError('Electron concentration cannot be negative in a semiconductor' ) elif hole_conc < 0: raise ValueError('Hole concentration cannot be negative in a semiconductor' ) elif intrinsic_conc < 0: raise ValueError( 'Intrinsic concentration cannot be negative in a semiconductor' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
96
'''simple docstring''' import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": lowerCAmelCase: Any = argparse.ArgumentParser() parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument( '--txt2img_unclip', default='kakaobrain/karlo-v1-alpha', type=str, required=False, help='The pretrained txt2img unclip.', ) lowerCAmelCase: Optional[int] = parser.parse_args() lowerCAmelCase: List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) lowerCAmelCase: Optional[Any] = CLIPImageProcessor() lowerCAmelCase: Tuple = CLIPVisionModelWithProjection.from_pretrained('openai/clip-vit-large-patch14') lowerCAmelCase: List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
96
1
'''simple docstring''' import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } UpperCAmelCase_ = { 'vocab_file': {'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'}, 'merges_file': {'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'}, 'tokenizer_config_file': { 'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json' }, } UpperCAmelCase_ = {'facebook/blenderbot-3B': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def _UpperCamelCase ( ): '''simple docstring''' UpperCAmelCase__ = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) UpperCAmelCase__ = bs[:] UpperCAmelCase__ = 0 for b in range(2**8 ): if b not in bs: bs.append(snake_case__ ) cs.append(2**8 + n ) n += 1 UpperCAmelCase__ = [chr(snake_case__ ) for n in cs] return dict(zip(snake_case__ , snake_case__ ) ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[Any] ): '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' lowerCAmelCase_ : Dict = VOCAB_FILES_NAMES lowerCAmelCase_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : Any , _UpperCAmelCase : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int="replace" , _UpperCAmelCase : List[str]="<s>" , _UpperCAmelCase : List[Any]="</s>" , _UpperCAmelCase : Optional[Any]="</s>" , _UpperCAmelCase : List[str]="<s>" , _UpperCAmelCase : int="<unk>" , _UpperCAmelCase : str="<pad>" , _UpperCAmelCase : Union[str, Any]="<mask>" , _UpperCAmelCase : int=False , **_UpperCAmelCase : str , ): """simple docstring""" UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else bos_token UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else eos_token UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else sep_token UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else cls_token UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else unk_token UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else mask_token super().__init__( errors=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , sep_token=A_ , cls_token=A_ , pad_token=A_ , mask_token=A_ , add_prefix_space=A_ , **A_ , ) with open(A_ , encoding="""utf-8""" ) as vocab_handle: UpperCAmelCase__ = json.load(A_ ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} UpperCAmelCase__ = errors # how to handle errors in decoding UpperCAmelCase__ = bytes_to_unicode() UpperCAmelCase__ = {v: k for k, v in self.byte_encoder.items()} with open(A_ , encoding="""utf-8""" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("""\n""" )[1:-1] UpperCAmelCase__ = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase__ = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCAmelCase__ = {} UpperCAmelCase__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase__ = re.compile(r"""\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return len(self.encoder ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE__ ( self : int , _UpperCAmelCase : Union[str, Any] ): """simple docstring""" if token in self.cache: return self.cache[token] UpperCAmelCase__ = tuple(A_ ) UpperCAmelCase__ = get_pairs(A_ ) if not pairs: return token while True: UpperCAmelCase__ = min(A_ , key=lambda _UpperCAmelCase : self.bpe_ranks.get(A_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(A_ ): try: UpperCAmelCase__ = word.index(A_ , A_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(A_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(A_ ) UpperCAmelCase__ = new_word if len(A_ ) == 1: break else: UpperCAmelCase__ = get_pairs(A_ ) UpperCAmelCase__ = """ """.join(A_ ) UpperCAmelCase__ = word return word def SCREAMING_SNAKE_CASE__ ( self : List[str] , _UpperCAmelCase : int ): """simple docstring""" UpperCAmelCase__ = [] for token in re.findall(self.pat , A_ ): UpperCAmelCase__ = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(A_ ).split(""" """ ) ) return bpe_tokens def SCREAMING_SNAKE_CASE__ ( self : List[str] , _UpperCAmelCase : Dict ): """simple docstring""" return self.encoder.get(A_ , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : List[Any] ): """simple docstring""" return self.decoder.get(A_ ) def SCREAMING_SNAKE_CASE__ ( self : Dict , _UpperCAmelCase : List[Any] ): """simple docstring""" UpperCAmelCase__ = """""".join(A_ ) UpperCAmelCase__ = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def SCREAMING_SNAKE_CASE__ ( self : str , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ): """simple docstring""" if not os.path.isdir(A_ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( A_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ = os.path.join( A_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(A_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=A_ , ensure_ascii=A_ ) + """\n""" ) UpperCAmelCase__ = 0 with open(A_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) UpperCAmelCase__ = token_index writer.write(""" """.join(A_ ) + """\n""" ) index += 1 return vocab_file, merge_file def SCREAMING_SNAKE_CASE__ ( self : List[str] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None , _UpperCAmelCase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is None: return [1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1, 1] + ([0] * len(A_ )) + [1] def SCREAMING_SNAKE_CASE__ ( self : Tuple , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE__ ( self : Tuple , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any]=False , **_UpperCAmelCase : Tuple ): """simple docstring""" UpperCAmelCase__ = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(A_ ) > 0 and not text[0].isspace()): UpperCAmelCase__ = """ """ + text return (text, kwargs) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): """simple docstring""" return token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , _UpperCAmelCase : "Conversation" ): """simple docstring""" UpperCAmelCase__ = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text ) else: # Generated responses should contain them already. inputs.append(A_ ) UpperCAmelCase__ = """ """.join(A_ ) UpperCAmelCase__ = self.encode(A_ ) if len(A_ ) > self.model_max_length: UpperCAmelCase__ = input_ids[-self.model_max_length :] logger.warning(f'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''' ) return input_ids
346
"""simple docstring""" from string import ascii_uppercase _lowercase = {char: i for i, char in enumerate(ascii_uppercase)} _lowercase = dict(enumerate(ascii_uppercase)) def _snake_case ( snake_case__ : str , snake_case__ : str ): A = len(snake_case__ ) A = 0 while True: if x == i: A = 0 if len(snake_case__ ) == len(snake_case__ ): break key += key[i] i += 1 return key def _snake_case ( snake_case__ : str , snake_case__ : str ): A = '' A = 0 for letter in message: if letter == " ": cipher_text += " " else: A = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def _snake_case ( snake_case__ : str , snake_case__ : str ): A = '' A = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: A = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def _snake_case ( ): A = 'THE GERMAN ATTACK' A = 'SECRET' A = generate_key(snake_case__ , snake_case__ ) A = cipher_text(snake_case__ , snake_case__ ) print(F'Encrypted Text = {s}' ) print(F'Original Text = {original_text(snake_case__ , snake_case__ )}' ) if __name__ == "__main__": import doctest doctest.testmod() main()
74
0
"""simple docstring""" from __future__ import annotations def __lowercase ( snake_case_ : list[list[int]] ) ->bool: '''simple docstring''' __A : Dict = len(snake_case_ ) # We need to create solution object to save path. __A : int = [[0 for _ in range(snake_case_ )] for _ in range(snake_case_ )] __A : str = run_maze(snake_case_ ,0 ,0 ,snake_case_ ) if solved: print('''\n'''.join(str(snake_case_ ) for row in solutions ) ) else: print('''No solution exists!''' ) return solved def __lowercase ( snake_case_ : list[list[int]] ,snake_case_ : int ,snake_case_ : int ,snake_case_ : list[list[int]] ) ->bool: '''simple docstring''' __A : str = len(snake_case_ ) # Final check point. if i == j == (size - 1): __A : Optional[Any] = 1 return True __A : int = (not i < 0) and (not j < 0) # Check lower bounds __A : List[str] = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. __A : str = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited __A : int = 1 # check for directions if ( run_maze(snake_case_ ,i + 1 ,snake_case_ ,snake_case_ ) or run_maze(snake_case_ ,snake_case_ ,j + 1 ,snake_case_ ) or run_maze(snake_case_ ,i - 1 ,snake_case_ ,snake_case_ ) or run_maze(snake_case_ ,snake_case_ ,j - 1 ,snake_case_ ) ): return True __A : Any = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
351
"""simple docstring""" a_ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) a_ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def __lowercase ( snake_case_ : float ,snake_case_ : str ,snake_case_ : str ) ->float: '''simple docstring''' __A : Tuple = from_type.lower().strip('''s''' ) __A : Optional[int] = to_type.lower().strip('''s''' ) __A : List[str] = UNIT_SYMBOL.get(snake_case_ ,snake_case_ ) __A : Any = UNIT_SYMBOL.get(snake_case_ ,snake_case_ ) if from_sanitized not in METRIC_CONVERSION: __A : int = ( F"""Invalid 'from_type' value: {from_type!r}.\n""" F"""Conversion abbreviations are: {', '.join(snake_case_ )}""" ) raise ValueError(snake_case_ ) if to_sanitized not in METRIC_CONVERSION: __A : str = ( F"""Invalid 'to_type' value: {to_type!r}.\n""" F"""Conversion abbreviations are: {', '.join(snake_case_ )}""" ) raise ValueError(snake_case_ ) __A : Optional[Any] = METRIC_CONVERSION[from_sanitized] __A : Optional[int] = METRIC_CONVERSION[to_sanitized] __A : Union[str, Any] = 1 if from_exponent > to_exponent: __A : Dict = from_exponent - to_exponent else: __A : Union[str, Any] = -(to_exponent - from_exponent) return value * pow(10 ,snake_case_ ) if __name__ == "__main__": from doctest import testmod testmod()
291
0
'''simple docstring''' import argparse import os import torch from transformers.utils import WEIGHTS_NAME UpperCamelCase_ = ["""small""", """medium""", """large"""] UpperCamelCase_ = """lm_head.decoder.weight""" UpperCamelCase_ = """lm_head.weight""" def _UpperCAmelCase ( _lowerCamelCase : str , _lowerCamelCase : str ) -> List[Any]: _lowerCAmelCase : Tuple = torch.load(_lowerCamelCase ) _lowerCAmelCase : Tuple = d.pop(_lowerCamelCase ) os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) torch.save(_lowerCamelCase , os.path.join(_lowerCamelCase , _lowerCamelCase ) ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument("""--dialogpt_path""", default=""".""", type=str) UpperCamelCase_ = parser.parse_args() for MODEL in DIALOGPT_MODELS: UpperCamelCase_ = os.path.join(args.dialogpt_path, F'{MODEL}_ft.pkl') UpperCamelCase_ = F'./DialoGPT-{MODEL}' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
309
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = """▁""" UpperCamelCase_ = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", """tokenizer_config_file""": """tokenizer_config.json""", } UpperCamelCase_ = { """vocab_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json""", }, """spm_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_config_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json""", }, } UpperCamelCase_ = { """facebook/m2m100_418M""": 10_24, } # fmt: off UpperCamelCase_ = { """m2m100""": ["""af""", """am""", """ar""", """ast""", """az""", """ba""", """be""", """bg""", """bn""", """br""", """bs""", """ca""", """ceb""", """cs""", """cy""", """da""", """de""", """el""", """en""", """es""", """et""", """fa""", """ff""", """fi""", """fr""", """fy""", """ga""", """gd""", """gl""", """gu""", """ha""", """he""", """hi""", """hr""", """ht""", """hu""", """hy""", """id""", """ig""", """ilo""", """is""", """it""", """ja""", """jv""", """ka""", """kk""", """km""", """kn""", """ko""", """lb""", """lg""", """ln""", """lo""", """lt""", """lv""", """mg""", """mk""", """ml""", """mn""", """mr""", """ms""", """my""", """ne""", """nl""", """no""", """ns""", """oc""", """or""", """pa""", """pl""", """ps""", """pt""", """ro""", """ru""", """sd""", """si""", """sk""", """sl""", """so""", """sq""", """sr""", """ss""", """su""", """sv""", """sw""", """ta""", """th""", """tl""", """tn""", """tr""", """uk""", """ur""", """uz""", """vi""", """wo""", """xh""", """yi""", """yo""", """zh""", """zu"""], """wmt21""": ["""en""", """ha""", """is""", """ja""", """cs""", """ru""", """zh""", """de"""] } class a_ (_a ): __lowerCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __lowerCAmelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase : Dict = ["""input_ids""", """attention_mask"""] __lowerCAmelCase : List[int] = [] __lowerCAmelCase : List[int] = [] def __init__( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=None , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<pad>" , snake_case_="<unk>" , snake_case_="m2m100" , snake_case_ = None , snake_case_=8 , **snake_case_ , ): _lowerCAmelCase : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : Optional[Any] = language_codes _lowerCAmelCase : Tuple = FAIRSEQ_LANGUAGE_CODES[language_codes] _lowerCAmelCase : str = {lang_code: f'__{lang_code}__' for lang_code in fairseq_language_code} _lowerCAmelCase : int = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(snake_case_ ) for lang_code in fairseq_language_code if self.get_lang_token(snake_case_ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=snake_case_ , tgt_lang=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , sep_token=snake_case_ , unk_token=snake_case_ , pad_token=snake_case_ , language_codes=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=snake_case_ , **snake_case_ , ) _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Any = load_json(snake_case_ ) _lowerCAmelCase : str = {v: k for k, v in self.encoder.items()} _lowerCAmelCase : Union[str, Any] = spm_file _lowerCAmelCase : Tuple = load_spm(snake_case_ , self.sp_model_kwargs ) _lowerCAmelCase : int = len(self.encoder ) _lowerCAmelCase : Union[str, Any] = { self.get_lang_token(snake_case_ ): self.encoder_size + i for i, lang_code in enumerate(snake_case_ ) } _lowerCAmelCase : List[str] = {lang_code: self.encoder_size + i for i, lang_code in enumerate(snake_case_ )} _lowerCAmelCase : Optional[Any] = {v: k for k, v in self.lang_token_to_id.items()} _lowerCAmelCase : Any = src_lang if src_lang is not None else """en""" _lowerCAmelCase : Optional[int] = tgt_lang _lowerCAmelCase : Tuple = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) _lowerCAmelCase : List[Any] = num_madeup_words @property def __UpperCamelCase ( self ): return len(self.encoder ) + len(self.lang_token_to_id ) @property def __UpperCamelCase ( self ): return self._src_lang @src_lang.setter def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __UpperCamelCase ( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def __UpperCamelCase ( self , snake_case_ ): if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(snake_case_ , self.encoder[self.unk_token] ) def __UpperCamelCase ( self , snake_case_ ): if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(snake_case_ , self.unk_token ) def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Any = [] _lowerCAmelCase : Optional[int] = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token _lowerCAmelCase : Optional[Any] = [] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def __UpperCamelCase ( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) _lowerCAmelCase : List[Any] = [1] * len(self.prefix_tokens ) _lowerCAmelCase : Dict = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(snake_case_ )) + suffix_ones return prefix_ones + ([0] * len(snake_case_ )) + ([0] * len(snake_case_ )) + suffix_ones def __UpperCamelCase ( self , snake_case_ , snake_case_ = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __UpperCamelCase ( self ): _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): _lowerCAmelCase : int = self.__dict__.copy() _lowerCAmelCase : str = None return state def __setstate__( self , snake_case_ ): _lowerCAmelCase : List[str] = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _lowerCAmelCase : str = {} _lowerCAmelCase : str = load_spm(self.spm_file , self.sp_model_kwargs ) def __UpperCamelCase ( self , snake_case_ , snake_case_ = None ): _lowerCAmelCase : Dict = Path(snake_case_ ) if not save_dir.is_dir(): raise OSError(f'{save_directory} should be a directory' ) _lowerCAmelCase : Any = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""vocab_file"""] ) _lowerCAmelCase : Any = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""spm_file"""] ) save_json(self.encoder , snake_case_ ) if os.path.abspath(self.spm_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , snake_case_ ) elif not os.path.isfile(self.spm_file ): with open(snake_case_ , """wb""" ) as fi: _lowerCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (str(snake_case_ ), str(snake_case_ )) def __UpperCamelCase ( self , snake_case_ , snake_case_ = "en" , snake_case_ = None , snake_case_ = "ro" , **snake_case_ , ): _lowerCAmelCase : Union[str, Any] = src_lang _lowerCAmelCase : Optional[Any] = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(snake_case_ , snake_case_ , **snake_case_ ) def __UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ): if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _lowerCAmelCase : Dict = src_lang _lowerCAmelCase : str = self(snake_case_ , add_special_tokens=snake_case_ , **snake_case_ ) _lowerCAmelCase : Union[str, Any] = self.get_lang_id(snake_case_ ) _lowerCAmelCase : Tuple = tgt_lang_id return inputs def __UpperCamelCase ( self ): self.set_src_lang_special_tokens(self.src_lang ) def __UpperCamelCase ( self ): self.set_tgt_lang_special_tokens(self.tgt_lang ) def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Optional[Any] = self.get_lang_token(snake_case_ ) _lowerCAmelCase : List[Any] = self.lang_token_to_id[lang_token] _lowerCAmelCase : Any = [self.cur_lang_id] _lowerCAmelCase : Any = [self.eos_token_id] def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Any = self.get_lang_token(snake_case_ ) _lowerCAmelCase : int = self.lang_token_to_id[lang_token] _lowerCAmelCase : str = [self.cur_lang_id] _lowerCAmelCase : str = [self.eos_token_id] def __UpperCamelCase ( self , snake_case_ ): return self.lang_code_to_token[lang] def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : List[str] = self.get_lang_token(snake_case_ ) return self.lang_token_to_id[lang_token] def _UpperCAmelCase ( _lowerCamelCase : str , _lowerCamelCase : Dict[str, Any] ) -> sentencepiece.SentencePieceProcessor: _lowerCAmelCase : Optional[Any] = sentencepiece.SentencePieceProcessor(**_lowerCamelCase ) spm.Load(str(_lowerCamelCase ) ) return spm def _UpperCAmelCase ( _lowerCamelCase : str ) -> Union[Dict, List]: with open(_lowerCamelCase , """r""" ) as f: return json.load(_lowerCamelCase ) def _UpperCAmelCase ( _lowerCamelCase : Tuple , _lowerCamelCase : str ) -> None: with open(_lowerCamelCase , """w""" ) as f: json.dump(_lowerCamelCase , _lowerCamelCase , indent=2 )
309
1
from math import factorial __A : Tuple = {str(d): factorial(d) for d in range(10)} def __UpperCamelCase ( _A : str ) ->int: """simple docstring""" return sum(DIGIT_FACTORIAL[d] for d in str(a__ ) ) def __UpperCamelCase ( ) ->int: """simple docstring""" lowerCamelCase_ =7 * factorial(9 ) + 1 return sum(i for i in range(3 , a__ ) if sum_of_digit_factorial(a__ ) == i ) if __name__ == "__main__": print(F"""{solution() = }""")
368
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __A : List[Any] = logging.get_logger(__name__) __A : List[Any] = [ ['attention', 'attn'], ['encoder_attention', 'encoder_attn'], ['q_lin', 'q_proj'], ['k_lin', 'k_proj'], ['v_lin', 'v_proj'], ['out_lin', 'out_proj'], ['norm_embeddings', 'layernorm_embedding'], ['position_embeddings', 'embed_positions'], ['embeddings', 'embed_tokens'], ['ffn.lin', 'fc'], ] def __UpperCamelCase ( _A : Optional[int] ) ->List[str]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: lowerCamelCase_ =k.replace(_A , _A ) if k.startswith("""encoder""" ): lowerCamelCase_ =k.replace(""".attn""" , """.self_attn""" ) lowerCamelCase_ =k.replace("""norm1""" , """self_attn_layer_norm""" ) lowerCamelCase_ =k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): lowerCamelCase_ =k.replace("""norm1""" , """self_attn_layer_norm""" ) lowerCamelCase_ =k.replace("""norm2""" , """encoder_attn_layer_norm""" ) lowerCamelCase_ =k.replace("""norm3""" , """final_layer_norm""" ) return k def __UpperCamelCase ( _A : Union[str, Any] ) ->Optional[int]: """simple docstring""" lowerCamelCase_ =[ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: lowerCamelCase_ =sd.pop(_A ) lowerCamelCase_ =k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd lowerCamelCase_ =v __A : Any = ['START'] @torch.no_grad() def __UpperCamelCase ( _A : List[Any] , _A : Union[str, Any] , _A : List[str] ) ->List[str]: """simple docstring""" lowerCamelCase_ =torch.load(_A , map_location="""cpu""" ) lowerCamelCase_ =model["""model"""] lowerCamelCase_ =BlenderbotConfig.from_json_file(_A ) lowerCamelCase_ =BlenderbotForConditionalGeneration(_A ) lowerCamelCase_ =m.model.state_dict().keys() lowerCamelCase_ =[] lowerCamelCase_ ={} for k, v in sd.items(): if k in IGNORE_KEYS: continue lowerCamelCase_ =rename_state_dict_key(_A ) if new_k not in valid_keys: failures.append([k, new_k] ) else: lowerCamelCase_ =v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(_A ) m.model.load_state_dict(_A , strict=_A ) m.half() m.save_pretrained(_A ) if __name__ == "__main__": __A : Any = argparse.ArgumentParser() # Required parameters parser.add_argument('--src_path', type=str, help='like blenderbot-model.bin') parser.add_argument('--save_dir', default='hf_blenderbot', type=str, help='Where to save converted model.') parser.add_argument( '--hf_config_json', default='blenderbot-3b-config.json', type=str, help='Path to config to use' ) __A : str = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
49
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase : Optional[int] = {"""configuration_fnet""": ["""FNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FNetConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Union[str, Any] = ["""FNetTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[str] = ["""FNetTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : str = [ """FNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FNetForMaskedLM""", """FNetForMultipleChoice""", """FNetForNextSentencePrediction""", """FNetForPreTraining""", """FNetForQuestionAnswering""", """FNetForSequenceClassification""", """FNetForTokenClassification""", """FNetLayer""", """FNetModel""", """FNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys lowercase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
20
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )-> List[str]: UpperCamelCase = 1.5 UpperCamelCase = int(factor * num_class_images ) UpperCamelCase = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=__UpperCamelCase , aesthetic_weight=0.1 ) os.makedirs(F"{class_data_dir}/images" , exist_ok=__UpperCamelCase ) if len(list(Path(F"{class_data_dir}/images" ).iterdir() ) ) >= num_class_images: return while True: UpperCamelCase = client.query(text=__UpperCamelCase ) if len(__UpperCamelCase ) >= factor * num_class_images or num_images > 1E4: break else: UpperCamelCase = int(factor * num_images ) UpperCamelCase = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=__UpperCamelCase , aesthetic_weight=0.1 , ) UpperCamelCase = 0 UpperCamelCase = 0 UpperCamelCase = tqdm(desc="""downloading real regularization images""" , total=__UpperCamelCase ) with open(F"{class_data_dir}/caption.txt" , """w""" ) as fa, open(F"{class_data_dir}/urls.txt" , """w""" ) as fa, open( F"{class_data_dir}/images.txt" , """w""" ) as fa: while total < num_class_images: UpperCamelCase = class_images[count] count += 1 try: UpperCamelCase = requests.get(images["""url"""] ) if img.status_code == 200: UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(F"{class_data_dir}/images/{total}.jpg" , """wb""" ) as f: f.write(img.content ) fa.write(images["""caption"""] + """\n""" ) fa.write(images["""url"""] + """\n""" ) fa.write(F"{class_data_dir}/images/{total}.jpg" + """\n""" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowercase__ ( )-> str: UpperCamelCase = argparse.ArgumentParser("""""" , add_help=__UpperCamelCase ) parser.add_argument("""--class_prompt""" , help="""text prompt to retrieve images""" , required=__UpperCamelCase , type=__UpperCamelCase ) parser.add_argument("""--class_data_dir""" , help="""path to save images""" , required=__UpperCamelCase , type=__UpperCamelCase ) parser.add_argument("""--num_class_images""" , help="""number of images to download""" , default=200 , type=__UpperCamelCase ) return parser.parse_args() if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
321
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { 'configuration_timesformer': ['TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimesformerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ 'TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimesformerModel', 'TimesformerForVideoClassification', 'TimesformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : int = args.pruning_method _snake_case : List[Any] = args.threshold _snake_case : Optional[Any] = args.model_name_or_path.rstrip("/" ) _snake_case : List[str] = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) _snake_case : List[Any] = torch.load(os.path.join(__lowercase , "pytorch_model.bin" ) ) _snake_case : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: _snake_case : Tuple = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: _snake_case : Optional[int] = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: _snake_case : List[Any] = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": _snake_case : Tuple = MagnitudeBinarizer.apply(inputs=__lowercase , threshold=__lowercase ) _snake_case : List[str] = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue _snake_case : Optional[Any] = name[:-6] _snake_case : Any = model[F"""{prefix_}mask_scores"""] _snake_case : Tuple = TopKBinarizer.apply(__lowercase , __lowercase ) _snake_case : Optional[Any] = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue _snake_case : int = name[:-6] _snake_case : List[Any] = model[F"""{prefix_}mask_scores"""] _snake_case : List[str] = ThresholdBinarizer.apply(__lowercase , __lowercase , __lowercase ) _snake_case : List[str] = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue _snake_case : int = name[:-6] _snake_case : Any = model[F"""{prefix_}mask_scores"""] _snake_case ,_snake_case : Union[str, Any] = -0.1, 1.1 _snake_case : Dict = torch.sigmoid(__lowercase ) _snake_case : List[str] = s * (r - l) + l _snake_case : Tuple = s_bar.clamp(min=0.0 , max=1.0 ) _snake_case : Union[str, Any] = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: _snake_case : Any = os.path.join( os.path.dirname(__lowercase ) , F"""bertarized_{os.path.basename(__lowercase )}""" ) if not os.path.isdir(__lowercase ): shutil.copytree(__lowercase , __lowercase ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(__lowercase , os.path.join(__lowercase , "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) __SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() main(args)
284
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { '''facebook/timesformer''': '''https://huggingface.co/facebook/timesformer/resolve/main/config.json''', } class __lowerCamelCase ( lowerCamelCase_): """simple docstring""" UpperCamelCase__ = """timesformer""" def __init__( self , UpperCAmelCase=224 , UpperCAmelCase=16 , UpperCAmelCase=3 , UpperCAmelCase=8 , UpperCAmelCase=768 , UpperCAmelCase=12 , UpperCAmelCase=12 , UpperCAmelCase=3072 , UpperCAmelCase="gelu" , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-6 , UpperCAmelCase=True , UpperCAmelCase="divided_space_time" , UpperCAmelCase=0 , **UpperCAmelCase , ): """simple docstring""" super().__init__(**UpperCAmelCase ) _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = num_frames _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = initializer_range _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = qkv_bias _UpperCAmelCase = attention_type _UpperCAmelCase = drop_path_rate
39
"""simple docstring""" from typing import Any class lowerCamelCase__ : def __init__( self , SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case : Tuple = data snake_case : Union[str, Any] = None class lowerCamelCase__ : def __init__( self ): """simple docstring""" snake_case : List[Any] = None def lowerCamelCase_ ( self ): """simple docstring""" snake_case : str = self.head while temp is not None: print(temp.data , end=" " ) snake_case : Optional[Any] = temp.next print() def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case : Union[str, Any] = Node(SCREAMING_SNAKE_CASE ) snake_case : List[Any] = self.head snake_case : Optional[int] = new_node def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" if node_data_a == node_data_a: return else: snake_case : int = self.head while node_a is not None and node_a.data != node_data_a: snake_case : Optional[Any] = node_a.next snake_case : Tuple = self.head while node_a is not None and node_a.data != node_data_a: snake_case : Union[str, Any] = node_a.next if node_a is None or node_a is None: return snake_case , snake_case : int = node_a.data, node_a.data if __name__ == "__main__": __A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("After swapping") ll.print_list()
148
0
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class UpperCAmelCase_ ( _a): '''simple docstring''' __UpperCamelCase : torch.FloatTensor __UpperCamelCase : torch.FloatTensor class UpperCAmelCase_ ( _a, _a): '''simple docstring''' __UpperCamelCase : int = 1 @register_to_config def __init__( self , __SCREAMING_SNAKE_CASE = 2_000 , __SCREAMING_SNAKE_CASE = 0.15 , __SCREAMING_SNAKE_CASE = 0.01 , __SCREAMING_SNAKE_CASE = 1_348.0 , __SCREAMING_SNAKE_CASE = 1e-5 , __SCREAMING_SNAKE_CASE = 1 , ): """simple docstring""" UpperCamelCase : Optional[Any] = sigma_max # setable values UpperCamelCase : Union[str, Any] = None self.set_sigmas(_lowercase , _lowercase , _lowercase , _lowercase ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" return sample def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" UpperCamelCase : Tuple = sampling_eps if sampling_eps is not None else self.config.sampling_eps UpperCamelCase : Tuple = torch.linspace(1 , _lowercase , _lowercase , device=_lowercase ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" UpperCamelCase : Any = sigma_min if sigma_min is not None else self.config.sigma_min UpperCamelCase : Union[str, Any] = sigma_max if sigma_max is not None else self.config.sigma_max UpperCamelCase : Dict = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(_lowercase , _lowercase ) UpperCamelCase : Union[str, Any] = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) UpperCamelCase : Union[str, Any] = torch.exp(torch.linspace(math.log(_lowercase ) , math.log(_lowercase ) , _lowercase ) ) UpperCamelCase : List[str] = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) UpperCamelCase : int = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) UpperCamelCase : Any = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda UpperCamelCase : List[str] = timesteps.to(self.discrete_sigmas.device ) UpperCamelCase : List[str] = self.discrete_sigmas[timesteps].to(sample.device ) UpperCamelCase : int = self.get_adjacent_sigma(_lowercase , _lowercase ).to(sample.device ) UpperCamelCase : Any = torch.zeros_like(_lowercase ) UpperCamelCase : Tuple = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods UpperCamelCase : Optional[Any] = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): UpperCamelCase : Dict = diffusion.unsqueeze(-1 ) UpperCamelCase : Union[str, Any] = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of UpperCamelCase : Dict = randn_tensor( sample.shape , layout=sample.layout , generator=_lowercase , device=sample.device , dtype=sample.dtype ) UpperCamelCase : Any = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? UpperCamelCase : str = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=_lowercase , prev_sample_mean=_lowercase ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction UpperCamelCase : Union[str, Any] = randn_tensor(sample.shape , layout=sample.layout , generator=_lowercase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr UpperCamelCase : Optional[int] = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() UpperCamelCase : Dict = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() UpperCamelCase : List[str] = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 UpperCamelCase : str = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term UpperCamelCase : Optional[Any] = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): UpperCamelCase : Tuple = step_size.unsqueeze(-1 ) UpperCamelCase : Tuple = sample + step_size * model_output UpperCamelCase : Union[str, Any] = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowercase ) def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ): """simple docstring""" UpperCamelCase : List[Any] = timesteps.to(original_samples.device ) UpperCamelCase : Dict = self.discrete_sigmas.to(original_samples.device )[timesteps] UpperCamelCase : Any = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(_lowercase ) * sigmas[:, None, None, None] ) UpperCamelCase : List[str] = noise + original_samples return noisy_samples def __len__( self ): """simple docstring""" return self.config.num_train_timesteps
366
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase : Tuple = logging.get_logger(__name__) __UpperCAmelCase : Union[str, Any] = { "kssteven/ibert-roberta-base": "https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json", "kssteven/ibert-roberta-large": "https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json", "kssteven/ibert-roberta-large-mnli": ( "https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json" ), } class UpperCAmelCase_ ( _a): '''simple docstring''' __UpperCamelCase : List[Any] = "ibert" def __init__( self , __SCREAMING_SNAKE_CASE=30_522 , __SCREAMING_SNAKE_CASE=768 , __SCREAMING_SNAKE_CASE=12 , __SCREAMING_SNAKE_CASE=12 , __SCREAMING_SNAKE_CASE=3_072 , __SCREAMING_SNAKE_CASE="gelu" , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=512 , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE=0.02 , __SCREAMING_SNAKE_CASE=1e-12 , __SCREAMING_SNAKE_CASE=1 , __SCREAMING_SNAKE_CASE=0 , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE="absolute" , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE="none" , **__SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__(pad_token_id=__SCREAMING_SNAKE_CASE , bos_token_id=__SCREAMING_SNAKE_CASE , eos_token_id=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = vocab_size UpperCamelCase : Optional[int] = hidden_size UpperCamelCase : Tuple = num_hidden_layers UpperCamelCase : Optional[Any] = num_attention_heads UpperCamelCase : Dict = hidden_act UpperCamelCase : Union[str, Any] = intermediate_size UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Any = attention_probs_dropout_prob UpperCamelCase : Dict = max_position_embeddings UpperCamelCase : Union[str, Any] = type_vocab_size UpperCamelCase : Optional[Any] = initializer_range UpperCamelCase : Union[str, Any] = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : int = quant_mode UpperCamelCase : Any = force_dequant class UpperCAmelCase_ ( _a): '''simple docstring''' @property def _lowercase ( self ): """simple docstring""" if self.task == "multiple-choice": UpperCamelCase : int = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCamelCase : Optional[int] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
315
0
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __snake_case = logging.get_logger(__name__) __snake_case = { '''microsoft/deberta-v2-xlarge''': '''https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xxlarge''': '''https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json''' ), '''microsoft/deberta-v2-xxlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json''' ), } class lowercase ( A__ ): """simple docstring""" _a = 'deberta-v2' def __init__( self , UpperCamelCase_=128100 , UpperCamelCase_=1536 , UpperCamelCase_=24 , UpperCamelCase_=24 , UpperCamelCase_=6144 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=512 , UpperCamelCase_=0 , UpperCamelCase_=0.02 , UpperCamelCase_=1e-7 , UpperCamelCase_=False , UpperCamelCase_=-1 , UpperCamelCase_=0 , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=0 , UpperCamelCase_="gelu" , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(**UpperCamelCase_ ) UpperCamelCase__ :Optional[int] = hidden_size UpperCamelCase__ :Union[str, Any] = num_hidden_layers UpperCamelCase__ :List[str] = num_attention_heads UpperCamelCase__ :Optional[Any] = intermediate_size UpperCamelCase__ :int = hidden_act UpperCamelCase__ :Dict = hidden_dropout_prob UpperCamelCase__ :str = attention_probs_dropout_prob UpperCamelCase__ :Union[str, Any] = max_position_embeddings UpperCamelCase__ :Dict = type_vocab_size UpperCamelCase__ :Dict = initializer_range UpperCamelCase__ :Any = relative_attention UpperCamelCase__ :int = max_relative_positions UpperCamelCase__ :List[Any] = pad_token_id UpperCamelCase__ :List[Any] = position_biased_input # Backwards compatibility if type(UpperCamelCase_ ) == str: UpperCamelCase__ :Optional[Any] = [x.strip() for x in pos_att_type.lower().split('''|''' )] UpperCamelCase__ :List[str] = pos_att_type UpperCamelCase__ :Dict = vocab_size UpperCamelCase__ :Optional[Any] = layer_norm_eps UpperCamelCase__ :Tuple = kwargs.get('''pooler_hidden_size''' , UpperCamelCase_ ) UpperCamelCase__ :List[Any] = pooler_dropout UpperCamelCase__ :str = pooler_hidden_act class lowercase ( A__ ): """simple docstring""" @property def lowerCAmelCase__ ( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase__ :str = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCamelCase__ :Optional[Any] = {0: '''batch''', 1: '''sequence'''} if self._config.type_vocab_size > 0: return OrderedDict( [('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis)] ) else: return OrderedDict([('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis)] ) @property def lowerCAmelCase__ ( self ): '''simple docstring''' return 12 def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = False , UpperCamelCase_ = None , UpperCamelCase_ = 3 , UpperCamelCase_ = 40 , UpperCamelCase_ = 40 , UpperCamelCase_ = None , ): '''simple docstring''' UpperCamelCase__ :List[Any] = super().generate_dummy_inputs(preprocessor=UpperCamelCase_ , framework=UpperCamelCase_ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
97
'''simple docstring''' def _lowerCamelCase ( lowercase : int ) -> bool: _a = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
63
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __lowercase : List[str] = {'''configuration_plbart''': ['''PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PLBartConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase : int = ['''PLBartTokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase : List[str] = [ '''PLBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PLBartForCausalLM''', '''PLBartForConditionalGeneration''', '''PLBartForSequenceClassification''', '''PLBartModel''', '''PLBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys __lowercase : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
366
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): """simple docstring""" lowercase__ = LongformerTokenizer lowercase__ = True lowercase__ = LongformerTokenizerFast lowercase__ = True def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __UpperCamelCase =dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) __UpperCamelCase =['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __UpperCamelCase ={'''unk_token''': '''<unk>'''} __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCamelCase__ ) ) def UpperCAmelCase_ ( self : Optional[int] , **UpperCamelCase__ : str ) -> Dict: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[str] , **UpperCamelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[str] , UpperCamelCase__ : List[str] ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase ='''lower newer''' __UpperCamelCase ='''lower newer''' return input_text, output_text def UpperCAmelCase_ ( self : int ) -> List[Any]: '''simple docstring''' __UpperCamelCase =self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __UpperCamelCase ='''lower newer''' __UpperCamelCase =['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __UpperCamelCase =tokenizer.tokenize(UpperCamelCase__ ) # , add_prefix_space=True) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokens + [tokenizer.unk_token] __UpperCamelCase =[0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() self.assertListEqual(tokenizer.encode('''Hello world!''' , add_special_tokens=UpperCamelCase__ ) , [0, 31414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode('''Hello world! cécé herlolip 418''' , add_special_tokens=UpperCamelCase__ ) , [0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2] , ) @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =self.tokenizer_class.from_pretrained('''allenai/longformer-base-4096''' ) __UpperCamelCase =tokenizer.encode('''sequence builders''' , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode('''multi-sequence build''' , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode( '''sequence builders''' , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ , UpperCamelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def UpperCAmelCase_ ( self : int ) -> Dict: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() __UpperCamelCase ='''Encode this sequence.''' __UpperCamelCase =tokenizer.byte_encoder[''' '''.encode('''utf-8''' )[0]] # Testing encoder arguments __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) tokenizer.add_special_tokens({'''bos_token''': '''<s>'''} ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) # Testing spaces after special tokens __UpperCamelCase ='''<mask>''' tokenizer.add_special_tokens( {'''mask_token''': AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ )} ) # mask token has a left space __UpperCamelCase =tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) __UpperCamelCase ='''Encode <mask> sequence''' __UpperCamelCase ='''Encode <mask>sequence''' __UpperCamelCase =tokenizer.encode(UpperCamelCase__ ) __UpperCamelCase =encoded.index(UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ ) __UpperCamelCase =encoded.index(UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase_ ( self : int ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __UpperCamelCase =self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase =self.tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase ='''A, <mask> AllenNLP sentence.''' __UpperCamelCase =tokenizer_r.encode_plus(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ ) __UpperCamelCase =tokenizer_p.encode_plus(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['''token_type_ids'''] ) , sum(tokens_p['''token_type_ids'''] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['''attention_mask'''] ) / len(tokens_r['''attention_mask'''] ) , sum(tokens_p['''attention_mask'''] ) / len(tokens_p['''attention_mask'''] ) , ) __UpperCamelCase =tokenizer_r.convert_ids_to_tokens(tokens_r['''input_ids'''] ) __UpperCamelCase =tokenizer_p.convert_ids_to_tokens(tokens_p['''input_ids'''] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['''input_ids'''] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r['''input_ids'''] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( UpperCamelCase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) self.assertSequenceEqual( UpperCamelCase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __UpperCamelCase =json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['''add_prefix_space'''] , UpperCamelCase__ ) self.assertEqual(post_processor_state['''add_prefix_space'''] , UpperCamelCase__ ) self.assertEqual(post_processor_state['''trim_offsets'''] , UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[Any] ) -> int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __UpperCamelCase ='''hello''' # `hello` is a token in the vocabulary of `pretrained_name` __UpperCamelCase =f"""{text_of_1_token} {text_of_1_token}""" __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ) + 1, len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ) + 1, len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ), len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ), len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =f""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ) + 1, 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ), 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ), 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , )
85
0
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __lowerCamelCase : Dict = logging.getLogger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :List[str] = 'token-classification' def __init__( self , A_ ): '''simple docstring''' if type(A_ ) == dict: UpperCamelCase : int = Namespace(**A_ ) UpperCamelCase : List[Any] = import_module("tasks" ) try: UpperCamelCase : List[str] = getattr(A_ , hparams.task_type ) UpperCamelCase : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( F"""Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ F"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) UpperCamelCase : str = self.token_classification_task.get_labels(hparams.labels ) UpperCamelCase : List[Any] = CrossEntropyLoss().ignore_index super().__init__(A_ , len(self.labels ) , self.mode ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return self.model(**A_ ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": UpperCamelCase : Dict = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids UpperCamelCase : int = self(**A_ ) UpperCamelCase : List[Any] = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.hparams for mode in ["train", "dev", "test"]: UpperCamelCase : Dict = self._feature_file(A_ ) if os.path.exists(A_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , A_ ) UpperCamelCase : Dict = torch.load(A_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) UpperCamelCase : List[Any] = self.token_classification_task.read_examples_from_file(args.data_dir , A_ ) UpperCamelCase : int = self.token_classification_task.convert_examples_to_features( A_ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=A_ , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , A_ ) torch.save(A_ , A_ ) def __UpperCamelCase( self , A_ , A_ , A_ = False ): '''simple docstring''' UpperCamelCase : Optional[Any] = self._feature_file(A_ ) logger.info("Loading features from cached file %s" , A_ ) UpperCamelCase : Optional[int] = torch.load(A_ ) UpperCamelCase : Optional[int] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) UpperCamelCase : Optional[Any] = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: UpperCamelCase : Tuple = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: UpperCamelCase : List[str] = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) UpperCamelCase : int = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(A_ , A_ , A_ , A_ ) , batch_size=A_ ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' """Compute validation""" "" UpperCamelCase : str = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": UpperCamelCase : Optional[int] = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids UpperCamelCase : List[str] = self(**A_ ) UpperCamelCase , UpperCamelCase : List[Any] = outputs[:2] UpperCamelCase : int = logits.detach().cpu().numpy() UpperCamelCase : Dict = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = torch.stack([x["val_loss"] for x in outputs] ).mean() UpperCamelCase : Union[str, Any] = np.concatenate([x["pred"] for x in outputs] , axis=0 ) UpperCamelCase : int = np.argmax(A_ , axis=2 ) UpperCamelCase : Optional[int] = np.concatenate([x["target"] for x in outputs] , axis=0 ) UpperCamelCase : List[Any] = dict(enumerate(self.labels ) ) UpperCamelCase : Union[str, Any] = [[] for _ in range(out_label_ids.shape[0] )] UpperCamelCase : Optional[int] = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) UpperCamelCase : Dict = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(A_ , A_ ), "precision": precision_score(A_ , A_ ), "recall": recall_score(A_ , A_ ), "f1": fa_score(A_ , A_ ), } UpperCamelCase : List[str] = dict(results.items() ) UpperCamelCase : Union[str, Any] = results return ret, preds_list, out_label_list def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase : int = self._eval_end(A_ ) UpperCamelCase : Union[str, Any] = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase : str = self._eval_end(A_ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 UpperCamelCase : Optional[int] = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def __UpperCamelCase( A_ , A_ ): '''simple docstring''' BaseTransformer.add_model_specific_args(A_ , A_ ) parser.add_argument( "--task_type" , default="NER" , type=A_ , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=128 , type=A_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=A_ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=A_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": __lowerCamelCase : List[str] = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __lowerCamelCase : Tuple = NERTransformer.add_model_specific_args(parser, os.getcwd()) __lowerCamelCase : List[str] = parser.parse_args() __lowerCamelCase : Any = NERTransformer(args) __lowerCamelCase : Any = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __lowerCamelCase : Optional[int] = sorted(glob.glob(os.path.join(args.output_dir, """checkpoint-epoch=*.ckpt"""), recursive=True)) __lowerCamelCase : Dict = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
52
from __future__ import annotations def __UpperCAmelCase ( a_ , a_ , a_ , a_): # noqa: E741 while r - l > 1: snake_case_ = (l + r) // 2 if v[m] >= key: snake_case_ = m else: snake_case_ = m # noqa: E741 return r def __UpperCAmelCase ( a_): if len(a_) == 0: return 0 snake_case_ = [0] * len(a_) snake_case_ = 1 snake_case_ = v[0] for i in range(1 , len(a_)): if v[i] < tail[0]: snake_case_ = v[i] elif v[i] > tail[length - 1]: snake_case_ = v[i] length += 1 else: snake_case_ = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
178
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __snake_case ( unittest.TestCase ): def __init__( self : Optional[int] , _snake_case : Optional[Any] , _snake_case : Dict=7 , _snake_case : Optional[int]=3 , _snake_case : Tuple=18 , _snake_case : int=30 , _snake_case : Dict=400 , _snake_case : List[Any]=True , _snake_case : str=None , _snake_case : Optional[Any]=True , _snake_case : Tuple=None , _snake_case : int=True , ): """simple docstring""" UpperCAmelCase_ = size if size is not None else {'''shortest_edge''': 20} UpperCAmelCase_ = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = image_size UpperCAmelCase_ = min_resolution UpperCAmelCase_ = max_resolution UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = do_flip_channel_order def lowerCamelCase ( self : str): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class __snake_case ( a , unittest.TestCase ): UpperCAmelCase__ : List[str] = MobileViTImageProcessor if is_vision_available() else None def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = MobileViTImageProcessingTester(self) @property def lowerCamelCase ( self : Optional[int]): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_snake_case , '''do_resize''')) self.assertTrue(hasattr(_snake_case , '''size''')) self.assertTrue(hasattr(_snake_case , '''do_center_crop''')) self.assertTrue(hasattr(_snake_case , '''center_crop''')) self.assertTrue(hasattr(_snake_case , '''do_flip_channel_order''')) def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) UpperCAmelCase_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def lowerCamelCase ( self : List[str]): """simple docstring""" pass def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict) # create random PIL images UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , Image.Image) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(_snake_case , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , numpify=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , np.ndarray) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(_snake_case , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , torchify=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , torch.Tensor) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(_snake_case , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
7
import comet # From: unbabel-comet import torch import datasets snake_case_ : Tuple = datasets.logging.get_logger(__name__) snake_case_ : str = "\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unbabel's Participation in the WMT20 Metrics Shared Task},\n booktitle = {Proceedings of the Fifth Conference on Machine Translation},\n month = {November},\n year = {2020},\n address = {Online},\n publisher = {Association for Computational Linguistics},\n pages = {909--918},\n}\n@inproceedings{rei-etal-2020-comet,\n title = \"{COMET}: A Neural Framework for {MT} Evaluation\",\n author = \"Rei, Ricardo and\n Stewart, Craig and\n Farinha, Ana C and\n Lavie, Alon\",\n booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\",\n month = nov,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\",\n pages = \"2685--2702\",\n}\n" snake_case_ : Tuple = "\\nCrosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM).\nWith the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition.\n\nSee the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information.\n" snake_case_ : Optional[int] = "\nCOMET score.\n\nArgs:\n\n`sources` (list of str): Source sentences\n`predictions` (list of str): candidate translations\n`references` (list of str): reference translations\n`cuda` (bool): If set to True, runs COMET using GPU\n`show_progress` (bool): Shows progress\n`model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None.\n\nReturns:\n `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`.\n `scores`: List of scores.\n\nExamples:\n\n >>> comet_metric = datasets.load_metric('comet')\n >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use\n >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"]\n >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"]\n >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"]\n >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source)\n >>> print([round(v, 2) for v in results[\"scores\"]])\n [0.19, 0.92]\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): def lowerCamelCase ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://unbabel.github.io/COMET/html/index.html''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''sources''': datasets.Value('''string''' , id='''sequence'''), '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , codebase_urls=['''https://github.com/Unbabel/COMET'''] , reference_urls=[ '''https://github.com/Unbabel/COMET''', '''https://www.aclweb.org/anthology/2020.emnlp-main.213/''', '''http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6''', ] , ) def lowerCamelCase ( self : List[Any] , _snake_case : Optional[int]): """simple docstring""" if self.config_name == "default": UpperCAmelCase_ = comet.load_from_checkpoint(comet.download_model('''wmt20-comet-da''')) else: UpperCAmelCase_ = comet.load_from_checkpoint(comet.download_model(self.config_name)) def lowerCamelCase ( self : List[Any] , _snake_case : str , _snake_case : List[str] , _snake_case : Tuple , _snake_case : int=None , _snake_case : Optional[Any]=False): """simple docstring""" if gpus is None: UpperCAmelCase_ = 1 if torch.cuda.is_available() else 0 UpperCAmelCase_ = {'''src''': sources, '''mt''': predictions, '''ref''': references} UpperCAmelCase_ = [dict(zip(_snake_case , _snake_case)) for t in zip(*data.values())] UpperCAmelCase_ , UpperCAmelCase_ = self.scorer.predict(_snake_case , gpus=_snake_case , progress_bar=_snake_case) return {"mean_score": mean_score, "scores": scores}
7
1
'''simple docstring''' import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __SCREAMING_SNAKE_CASE :Any = 50000 __SCREAMING_SNAKE_CASE :List[str] = 5000 __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Dict = os.path.split(__file__) __SCREAMING_SNAKE_CASE :str = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def UpperCAmelCase_ ( __lowercase : datasets.Dataset , __lowercase : Any ) -> Dict: '''simple docstring''' for i in range(__lowercase ): _UpperCAmelCase = dataset[i] @get_duration def UpperCAmelCase_ ( __lowercase : datasets.Dataset , __lowercase : int , __lowercase : Union[str, Any] ) -> str: '''simple docstring''' for i in range(0 , len(__lowercase ) , __lowercase ): _UpperCAmelCase = dataset[i : i + batch_size] @get_duration def UpperCAmelCase_ ( __lowercase : datasets.Dataset , __lowercase : int , __lowercase : Union[str, Any] ) -> Dict: '''simple docstring''' with dataset.formatted_as(type=__lowercase ): for i in range(__lowercase ): _UpperCAmelCase = dataset[i] @get_duration def UpperCAmelCase_ ( __lowercase : datasets.Dataset , __lowercase : Optional[int] , __lowercase : List[str] , __lowercase : Any ) -> Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=__lowercase ): for i in range(0 , __lowercase , __lowercase ): _UpperCAmelCase = dataset[i : i + batch_size] def UpperCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' _UpperCAmelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCAmelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] _UpperCAmelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCAmelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCAmelCase = generate_example_dataset( os.path.join(__lowercase , "dataset.arrow" ) , __lowercase , num_examples=__lowercase , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(__lowercase ) ) _UpperCAmelCase = func(__lowercase , **__lowercase ) print("shuffling dataset" ) _UpperCAmelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(__lowercase ) ) _UpperCAmelCase = func( __lowercase , **__lowercase ) with open(__lowercase , "wb" ) as f: f.write(json.dumps(__lowercase ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
22
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name lowercase__ = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline\n >>> from diffusers.utils import load_image\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-prior\", torch_dtype=torch.float16\n ... )\n >>> pipe_prior.to(\"cuda\")\n\n >>> prompt = \"A red cartoon frog, 4k\"\n >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False)\n\n >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-decoder\", torch_dtype=torch.float16\n ... )\n >>> pipe.to(\"cuda\")\n\n >>> init_image = load_image(\n ... \"https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main\"\n ... \"/kandinsky/frog.png\"\n ... )\n\n >>> image = pipe(\n ... image=init_image,\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... strength=0.2,\n ... ).images\n\n >>> image[0].save(\"red_frog.png\")\n ```\n" def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=8 ): UpperCAmelCase : Tuple = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase : Union[str, Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_=5_12 , UpperCAmelCase_=5_12 ): UpperCAmelCase : List[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) UpperCAmelCase : Tuple = np.array(pil_image.convert('RGB' ) ) UpperCAmelCase : List[Any] = arr.astype(np.floataa ) / 127.5 - 1 UpperCAmelCase : List[str] = np.transpose(UpperCAmelCase_ , [2, 0, 1] ) UpperCAmelCase : Tuple = torch.from_numpy(UpperCAmelCase_ ).unsqueeze(0 ) return image class A_ ( _snake_case ): '''simple docstring''' def __init__( self : List[str] , lowercase_ : UNetaDConditionModel , lowercase_ : DDPMScheduler , lowercase_ : VQModel , ) -> str: super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase : Dict = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : str ) -> Dict: # get the original timestep using init_timestep UpperCAmelCase : Tuple = min(int(num_inference_steps * strength ) , lowercase_ ) UpperCAmelCase : Any = max(num_inference_steps - init_timestep , 0 ) UpperCAmelCase : Any = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self : str , lowercase_ : List[Any] , lowercase_ : Dict , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : Dict , lowercase_ : Dict , lowercase_ : Any=None ) -> Optional[Any]: if not isinstance(lowercase_ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase_ )}""" ) UpperCAmelCase : Any = image.to(device=lowercase_ , dtype=lowercase_ ) UpperCAmelCase : List[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: UpperCAmelCase : List[Any] = image else: if isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(lowercase_ )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(lowercase_ ) ] UpperCAmelCase : Dict = torch.cat(lowercase_ , dim=0 ) else: UpperCAmelCase : List[str] = self.movq.encode(lowercase_ ).latent_dist.sample(lowercase_ ) UpperCAmelCase : Optional[Any] = self.movq.config.scaling_factor * init_latents UpperCAmelCase : int = torch.cat([init_latents] , dim=0 ) UpperCAmelCase : Union[str, Any] = init_latents.shape UpperCAmelCase : Union[str, Any] = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) # get latents UpperCAmelCase : Union[str, Any] = self.scheduler.add_noise(lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase : Optional[Any] = init_latents return latents def UpperCAmelCase_ ( self : Union[str, Any] , lowercase_ : Union[str, Any]=0 ) -> Tuple: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) UpperCAmelCase : int = torch.device(f"""cuda:{gpu_id}""" ) UpperCAmelCase : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCAmelCase_ ( self : Union[str, Any] , lowercase_ : int=0 ) -> str: if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) UpperCAmelCase : int = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase : Any = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase : Dict = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase : Optional[Any] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self : Tuple ) -> List[str]: if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self : Optional[Any] , lowercase_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowercase_ : Union[torch.FloatTensor, PIL.Image.Image, List[torch.FloatTensor], List[PIL.Image.Image]] , lowercase_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowercase_ : int = 512 , lowercase_ : int = 512 , lowercase_ : int = 100 , lowercase_ : float = 4.0 , lowercase_ : float = 0.3 , lowercase_ : int = 1 , lowercase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowercase_ : Optional[str] = "pil" , lowercase_ : bool = True , ) -> Dict: UpperCAmelCase : str = self._execution_device UpperCAmelCase : int = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase : List[str] = image_embeds.shape[0] if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : str = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase : Dict = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase : Tuple = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) if not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : Optional[Any] = [image] if not all(isinstance(lowercase_ , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( f"""Input is in incorrect format: {[type(lowercase_ ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) UpperCAmelCase : Dict = torch.cat([prepare_image(lowercase_ , lowercase_ , lowercase_ ) for i in image] , dim=0 ) UpperCAmelCase : Optional[int] = image.to(dtype=image_embeds.dtype , device=lowercase_ ) UpperCAmelCase : List[str] = self.movq.encode(lowercase_ )['latents'] UpperCAmelCase : int = latents.repeat_interleave(lowercase_ , dim=0 ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase : Dict = self.get_timesteps(lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase : Optional[Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) UpperCAmelCase : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) UpperCAmelCase : int = self.prepare_latents( lowercase_ , lowercase_ , lowercase_ , lowercase_ , image_embeds.dtype , lowercase_ , lowercase_ ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase : int = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase : int = {'image_embeds': image_embeds} UpperCAmelCase : Optional[int] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase : List[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase : Optional[int] = noise_pred.chunk(2 ) UpperCAmelCase : int = variance_pred.chunk(2 ) UpperCAmelCase : Union[str, Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase : Dict = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase : Optional[Any] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase : List[Any] = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase : int = image * 0.5 + 0.5 UpperCAmelCase : List[Any] = image.clamp(0 , 1 ) UpperCAmelCase : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase : str = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
370
'''simple docstring''' def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): while a != 0: UpperCAmelCase , UpperCAmelCase : Tuple = b % a, a return b def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): if gcd(UpperCAmelCase_ , UpperCAmelCase_ ) != 1: UpperCAmelCase : List[str] = F"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(UpperCAmelCase_ ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Any = 1, 0, a UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Dict = 0, 1, m while va != 0: UpperCAmelCase : Tuple = ua // va UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[Any] = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
280
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : List[Any] = logging.get_logger(__name__) lowercase : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = 'pegasus' _A = ['past_key_values'] _A = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self :Dict , a :Dict=5_0_2_6_5 , a :Dict=1_0_2_4 , a :Union[str, Any]=1_2 , a :Any=4_0_9_6 , a :str=1_6 , a :str=1_2 , a :Optional[Any]=4_0_9_6 , a :int=1_6 , a :Optional[int]=0.0 , a :Optional[int]=0.0 , a :List[Any]=True , a :Union[str, Any]=True , a :int="gelu" , a :Dict=1_0_2_4 , a :List[Any]=0.1 , a :List[str]=0.0 , a :List[Any]=0.0 , a :str=0.02 , a :int=0 , a :Any=False , a :Dict=0 , a :int=1 , a :Optional[Any]=1 , **a :Optional[int] , ) -> str: __UpperCamelCase : List[Any] = vocab_size __UpperCamelCase : Union[str, Any] = max_position_embeddings __UpperCamelCase : str = d_model __UpperCamelCase : Dict = encoder_ffn_dim __UpperCamelCase : int = encoder_layers __UpperCamelCase : int = encoder_attention_heads __UpperCamelCase : List[Any] = decoder_ffn_dim __UpperCamelCase : List[Any] = decoder_layers __UpperCamelCase : List[str] = decoder_attention_heads __UpperCamelCase : str = dropout __UpperCamelCase : Union[str, Any] = attention_dropout __UpperCamelCase : List[str] = activation_dropout __UpperCamelCase : Optional[Any] = activation_function __UpperCamelCase : Tuple = init_std __UpperCamelCase : Optional[int] = encoder_layerdrop __UpperCamelCase : Union[str, Any] = decoder_layerdrop __UpperCamelCase : Optional[Any] = use_cache __UpperCamelCase : Union[str, Any] = encoder_layers __UpperCamelCase : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=a , eos_token_id=a , is_encoder_decoder=a , decoder_start_token_id=a , forced_eos_token_id=a , **a , ) @property def _lowerCamelCase ( self :Dict ) -> int: return self.encoder_attention_heads @property def _lowerCamelCase ( self :Optional[Any] ) -> int: return self.d_model
232
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase : List[str] = {'configuration_vit_msn': ['VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTMSNConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = [ 'VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTMSNModel', 'ViTMSNForImageClassification', 'ViTMSNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
232
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _UpperCamelCase : Tuple = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Optional[Any] = ["""PerceiverFeatureExtractor"""] _UpperCamelCase : Dict = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : int = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _UpperCamelCase : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
353
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCamelCase : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class snake_case ( UpperCAmelCase ): __magic_name__ = '''beit''' def __init__( self : int , A : int=8_1_9_2 , A : List[Any]=7_6_8 , A : str=1_2 , A : str=1_2 , A : Dict=3_0_7_2 , A : Optional[int]="gelu" , A : List[Any]=0.0 , A : Union[str, Any]=0.0 , A : Optional[Any]=0.02 , A : Optional[int]=1E-12 , A : Dict=2_2_4 , A : str=1_6 , A : Optional[Any]=3 , A : List[Any]=False , A : Union[str, Any]=False , A : Optional[Any]=False , A : int=False , A : List[str]=0.1 , A : Union[str, Any]=0.1 , A : str=True , A : Tuple=[3, 5, 7, 1_1] , A : List[str]=[1, 2, 3, 6] , A : Optional[Any]=True , A : Union[str, Any]=0.4 , A : Any=2_5_6 , A : List[Any]=1 , A : Optional[Any]=False , A : Any=2_5_5 , **A : List[Any] , ): '''simple docstring''' super().__init__(**A ) a : Optional[int] = vocab_size a : Dict = hidden_size a : Optional[int] = num_hidden_layers a : Tuple = num_attention_heads a : Optional[int] = intermediate_size a : Optional[Any] = hidden_act a : Optional[int] = hidden_dropout_prob a : Optional[int] = attention_probs_dropout_prob a : Optional[Any] = initializer_range a : Union[str, Any] = layer_norm_eps a : Union[str, Any] = image_size a : str = patch_size a : Optional[Any] = num_channels a : List[str] = use_mask_token a : Optional[Any] = use_absolute_position_embeddings a : Any = use_relative_position_bias a : Any = use_shared_relative_position_bias a : Dict = layer_scale_init_value a : Optional[int] = drop_path_rate a : Dict = use_mean_pooling # decode head attributes (semantic segmentation) a : Optional[Any] = out_indices a : List[str] = pool_scales # auxiliary head attributes (semantic segmentation) a : Tuple = use_auxiliary_head a : Dict = auxiliary_loss_weight a : Any = auxiliary_channels a : Dict = auxiliary_num_convs a : List[str] = auxiliary_concat_input a : List[Any] = semantic_loss_ignore_index class snake_case ( UpperCAmelCase ): __magic_name__ = version.parse('''1.11''' ) @property def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowerCamelCase__ ( self : Optional[Any] ): '''simple docstring''' return 1E-4
186
0
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( "stable diffusion controlnet", "0.22.0", "Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.", standard_warn=False, stacklevel=3, )
231
from ..utils import DummyObject, requires_backends class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Union[str, Any] , **__lowerCAmelCase : Dict ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Optional[int] , **__lowerCAmelCase : int ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : List[str] , **__lowerCAmelCase : int ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Union[str, Any] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Dict , **__lowerCAmelCase : Any ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Optional[int] , **__lowerCAmelCase : Dict ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : int , **__lowerCAmelCase : Any ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] )
231
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _UpperCamelCase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
335
import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowercase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' @register_to_config def __init__(self , *, __a = 4 , __a = 768 , __a , __a , ) -> str: """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Parameter(torch.zeros(__a ) ) # parameters for additional clip time embeddings UpperCAmelCase__ = nn.Linear(__a , __a ) UpperCAmelCase__ = nn.Linear(__a , __a ) # parameters for encoder hidden states UpperCAmelCase__ = clip_extra_context_tokens UpperCAmelCase__ = nn.Linear( __a , self.clip_extra_context_tokens * cross_attention_dim ) UpperCAmelCase__ = nn.Linear(__a , __a ) UpperCAmelCase__ = nn.LayerNorm(__a ) def UpperCamelCase__ (self , *, __a , __a , __a , __a ) -> Optional[Any]: """simple docstring""" if do_classifier_free_guidance: # Add the classifier free guidance embeddings to the image embeddings UpperCAmelCase__ = image_embeddings.shape[0] UpperCAmelCase__ = self.learned_classifier_free_guidance_embeddings.unsqueeze(0 ) UpperCAmelCase__ = classifier_free_guidance_embeddings.expand( __a , -1 ) UpperCAmelCase__ = torch.cat([classifier_free_guidance_embeddings, image_embeddings] , dim=0 ) # The image embeddings batch size and the text embeddings batch size are equal assert image_embeddings.shape[0] == prompt_embeds.shape[0] UpperCAmelCase__ = prompt_embeds.shape[0] # "Specifically, we modify the architecture described in Nichol et al. (2021) by projecting and # adding CLIP embeddings to the existing timestep embedding, ... UpperCAmelCase__ = self.embedding_proj(__a ) UpperCAmelCase__ = self.clip_image_embeddings_project_to_time_embeddings(__a ) UpperCAmelCase__ = time_projected_image_embeddings + time_projected_prompt_embeds # ... and by projecting CLIP embeddings into four # extra tokens of context that are concatenated to the sequence of outputs from the GLIDE text encoder" UpperCAmelCase__ = self.clip_extra_context_tokens_proj(__a ) UpperCAmelCase__ = clip_extra_context_tokens.reshape(__a , -1 , self.clip_extra_context_tokens ) UpperCAmelCase__ = clip_extra_context_tokens.permute(0 , 2 , 1 ) UpperCAmelCase__ = self.encoder_hidden_states_proj(__a ) UpperCAmelCase__ = self.text_encoder_hidden_states_norm(__a ) UpperCAmelCase__ = torch.cat([clip_extra_context_tokens, text_encoder_hidden_states] , dim=1 ) return text_encoder_hidden_states, additive_clip_time_embeddings
335
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): __magic_name__ = StableDiffusionInpaintPipeline __magic_name__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __magic_name__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __magic_name__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __magic_name__ = frozenset([] ) def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) A : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=__SCREAMING_SNAKE_CASE , ) A : List[str] = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE ) torch.manual_seed(0 ) A : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) A : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) A : List[str] = CLIPTextModel(__SCREAMING_SNAKE_CASE ) A : Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) A : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=0 ) -> int: """simple docstring""" A : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) A : Tuple = image.cpu().permute(0 , 2 , 3 , 1 )[0] A : Dict = Image.fromarray(np.uinta(__SCREAMING_SNAKE_CASE ) ).convert('''RGB''' ).resize((64, 64) ) A : List[Any] = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((64, 64) ) if str(__SCREAMING_SNAKE_CASE ).startswith('''mps''' ): A : Tuple = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: A : Dict = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) A : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator A : Optional[Any] = self.get_dummy_components() A : int = StableDiffusionInpaintPipeline(**__SCREAMING_SNAKE_CASE ) A : Any = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) A : Any = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) A : Optional[Any] = sd_pipe(**__SCREAMING_SNAKE_CASE ).images A : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A : Tuple = np.array([0.4_727, 0.5_735, 0.3_941, 0.5_446, 0.5_926, 0.4_394, 0.5_062, 0.4_654, 0.4_476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __lowerCAmelCase ( self ) -> str: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class A ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) A : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) A : List[str] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) A : List[str] = '''stabilityai/stable-diffusion-2-inpainting''' A : int = StableDiffusionInpaintPipeline.from_pretrained(__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() A : Union[str, Any] = '''Face of a yellow cat, high resolution, sitting on a park bench''' A : Optional[Any] = torch.manual_seed(0 ) A : str = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type='''np''' , ) A : Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9e-3 def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) A : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) A : int = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) A : Any = '''stabilityai/stable-diffusion-2-inpainting''' A : List[Any] = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , safety_checker=__SCREAMING_SNAKE_CASE , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() A : Any = '''Face of a yellow cat, high resolution, sitting on a park bench''' A : Any = torch.manual_seed(0 ) A : Optional[int] = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type='''np''' , ) A : Optional[int] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5e-1 def __lowerCAmelCase ( self ) -> Any: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() A : str = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) A : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) A : Dict = '''stabilityai/stable-diffusion-2-inpainting''' A : List[Any] = PNDMScheduler.from_pretrained(__SCREAMING_SNAKE_CASE , subfolder='''scheduler''' ) A : Dict = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() A : Dict = '''Face of a yellow cat, high resolution, sitting on a park bench''' A : Optional[Any] = torch.manual_seed(0 ) A : Optional[Any] = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''np''' , ) A : List[str] = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
3
'''simple docstring''' import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() UpperCAmelCase : Tuple = { 'bart': ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'bert': ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-base-cased-finetuned-mrpc': ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'dpr': ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'gpt2': ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlnet': ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm': ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm-roberta': ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'transfo-xl': ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'openai-gpt': ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'roberta': ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'layoutlm': ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'roberta-large-mnli': ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'camembert': ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'flaubert': ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert': ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert-base-distilled-squad': ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert': ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert-visual-feature-encoder': ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'ctrl': ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'albert': ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 't5': ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'electra': ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'wav2vec2': ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def a__ ( a__ , a__ , a__ , a__ , a__=False , a__=True ): """simple docstring""" if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.' ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: __SCREAMING_SNAKE_CASE = cached_file(a__ , a__ , force_download=not use_cached_models ) __SCREAMING_SNAKE_CASE = config_class.from_json_file(a__ ) __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True print(F'Building TensorFlow model from configuration: {config}' ) __SCREAMING_SNAKE_CASE = model_class(a__ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): __SCREAMING_SNAKE_CASE = cached_file( a__ , a__ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: __SCREAMING_SNAKE_CASE = load_pytorch_checkpoint_in_tfa_model(a__ , a__ ) if compare_with_pt_model: __SCREAMING_SNAKE_CASE = tf_model(tf_model.dummy_inputs , training=a__ ) # build the network __SCREAMING_SNAKE_CASE = torch.load(a__ , map_location="""cpu""" ) __SCREAMING_SNAKE_CASE = pt_model_class.from_pretrained( pretrained_model_name_or_path=a__ , config=a__ , state_dict=a__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = pt_model(**pt_model.dummy_inputs ) __SCREAMING_SNAKE_CASE = pto[0].numpy() __SCREAMING_SNAKE_CASE = tfo[0].numpy() __SCREAMING_SNAKE_CASE = np.amax(np.abs(np_pt - np_tf ) ) print(F'Max absolute difference between models outputs {diff}' ) assert diff <= 2E-2, F'Error, model absolute difference is >2e-2: {diff}' # Save pytorch-model print(F'Save TensorFlow model to {tf_dump_path}' ) tf_model.save_weights(a__ , save_format="""h5""" ) def a__ ( a__ , a__ , a__=None , a__=None , a__=False , a__=False , a__=False , a__=False , ): """simple docstring""" if args_model_type is None: __SCREAMING_SNAKE_CASE = list(MODEL_CLASSES.keys() ) else: __SCREAMING_SNAKE_CASE = [args_model_type] for j, model_type in enumerate(a__ , start=1 ): print("""=""" * 1_00 ) print(F' Converting model type {j}/{len(a__ )}: {model_type}' ) print("""=""" * 1_00 ) if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.' ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: __SCREAMING_SNAKE_CASE = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: __SCREAMING_SNAKE_CASE = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(a__ , a__ ) , start=1 ): print("""-""" * 1_00 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(F' Skipping finetuned checkpoint {model_shortcut_name}' ) continue __SCREAMING_SNAKE_CASE = model_shortcut_name elif only_convert_finetuned_models: print(F' Skipping not finetuned checkpoint {model_shortcut_name}' ) continue print( F' Converting checkpoint {i}/{len(a__ )}: {model_shortcut_name} - model_type {model_type}' ) print("""-""" * 1_00 ) if config_shortcut_name in aws_config_map: __SCREAMING_SNAKE_CASE = cached_file(a__ , a__ , force_download=not use_cached_models ) else: __SCREAMING_SNAKE_CASE = config_shortcut_name if model_shortcut_name in aws_model_maps: __SCREAMING_SNAKE_CASE = cached_file(a__ , a__ , force_download=not use_cached_models ) else: __SCREAMING_SNAKE_CASE = model_shortcut_name if os.path.isfile(a__ ): __SCREAMING_SNAKE_CASE = """converted_model""" convert_pt_checkpoint_to_tf( model_type=a__ , pytorch_checkpoint_path=a__ , config_file=a__ , tf_dump_path=os.path.join(a__ , model_shortcut_name + """-tf_model.h5""" ) , compare_with_pt_model=a__ , ) if remove_cached_files: os.remove(a__ ) os.remove(a__ ) if __name__ == "__main__": UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_dump_path', default=None, type=str, required=True, help='Path to the output Tensorflow dump file.' ) parser.add_argument( '--model_type', default=None, type=str, help=( f"""Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and """ 'convert all the models from AWS.' ), ) parser.add_argument( '--pytorch_checkpoint_path', default=None, type=str, help=( 'Path to the PyTorch checkpoint path or shortcut name to download from AWS. ' 'If not given, will download and convert all the checkpoints from AWS.' ), ) parser.add_argument( '--config_file', default=None, type=str, help=( 'The config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture. If not given and ' '--pytorch_checkpoint_path is not given or is a shortcut name ' 'use the configuration associated to the shortcut name on the AWS' ), ) parser.add_argument( '--compare_with_pt_model', action='store_true', help='Compare Tensorflow and PyTorch model predictions.' ) parser.add_argument( '--use_cached_models', action='store_true', help='Use cached models if possible instead of updating to latest checkpoint versions.', ) parser.add_argument( '--remove_cached_files', action='store_true', help='Remove pytorch models after conversion (save memory when converting in batches).', ) parser.add_argument('--only_convert_finetuned_models', action='store_true', help='Only convert finetuned models.') UpperCAmelCase : List[Any] = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
267
0
import math def lowerCamelCase_ ( _UpperCamelCase ) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase_ ( _UpperCamelCase = 10_001 ) -> int: """simple docstring""" try: snake_case_ : str = int(_UpperCamelCase ) except (TypeError, ValueError): raise TypeError('''Parameter nth must be int or castable to int.''' ) from None if nth <= 0: raise ValueError('''Parameter nth must be greater than or equal to one.''' ) snake_case_ : list[int] = [] snake_case_ : Union[str, Any] = 2 while len(_UpperCamelCase ) < nth: if is_prime(_UpperCamelCase ): primes.append(_UpperCamelCase ) num += 1 else: num += 1 return primes[len(_UpperCamelCase ) - 1] if __name__ == "__main__": print(F'''{solution() = }''')
279
def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''multiplicative_persistence() does not accept negative values''' ) snake_case_ : List[Any] = 0 snake_case_ : Tuple = str(_UpperCamelCase ) while len(_UpperCamelCase ) != 1: snake_case_ : Tuple = [int(_UpperCamelCase ) for i in num_string] snake_case_ : Dict = 1 for i in range(0 , len(_UpperCamelCase ) ): total *= numbers[i] snake_case_ : str = str(_UpperCamelCase ) steps += 1 return steps def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ): raise ValueError('''additive_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''additive_persistence() does not accept negative values''' ) snake_case_ : Any = 0 snake_case_ : Tuple = str(_UpperCamelCase ) while len(_UpperCamelCase ) != 1: snake_case_ : List[str] = [int(_UpperCamelCase ) for i in num_string] snake_case_ : Optional[int] = 0 for i in range(0 , len(_UpperCamelCase ) ): total += numbers[i] snake_case_ : Tuple = str(_UpperCamelCase ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
279
1
import heapq as hq import math from collections.abc import Iterator class __UpperCamelCase : """simple docstring""" def __init__( self : str , _A : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = str(id_ ) __SCREAMING_SNAKE_CASE : List[str] = None __SCREAMING_SNAKE_CASE : Optional[int] = None __SCREAMING_SNAKE_CASE : Tuple = [] __SCREAMING_SNAKE_CASE : Any = {} # {vertex:distance} def __lt__( self : Tuple , _A : Optional[int] ): """simple docstring""" return self.key < other.key def __repr__( self : List[Any] ): """simple docstring""" return self.id def UpperCAmelCase__ ( self : int , _A : str ): """simple docstring""" self.neighbors.append(A_ ) def UpperCAmelCase__ ( self : Dict , _A : List[Any] , _A : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = weight def a__ ( snake_case , snake_case , snake_case , snake_case ): """simple docstring""" # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __UpperCamelCase ) graph[b - 1].add_edge(graph[a - 1] , __UpperCamelCase ) def a__ ( snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = [] for u in graph: __SCREAMING_SNAKE_CASE : Union[str, Any] = math.inf __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Optional[Any] = 0 __SCREAMING_SNAKE_CASE : Tuple = graph[:] while q: __SCREAMING_SNAKE_CASE : Any = min(__UpperCamelCase ) q.remove(__UpperCamelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): __SCREAMING_SNAKE_CASE : Optional[Any] = u __SCREAMING_SNAKE_CASE : Any = u.edges[v.id] for i in range(1 , len(__UpperCamelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def a__ ( snake_case , snake_case ): """simple docstring""" for u in graph: __SCREAMING_SNAKE_CASE : List[Any] = math.inf __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Optional[Any] = 0 __SCREAMING_SNAKE_CASE : List[Any] = list(__UpperCamelCase ) hq.heapify(__UpperCamelCase ) while h: __SCREAMING_SNAKE_CASE : List[str] = hq.heappop(__UpperCamelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): __SCREAMING_SNAKE_CASE : Dict = u __SCREAMING_SNAKE_CASE : List[Any] = u.edges[v.id] hq.heapify(__UpperCamelCase ) for i in range(1 , len(__UpperCamelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def a__ ( ): """simple docstring""" pass if __name__ == "__main__": import doctest doctest.testmod()
303
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version A__ : Tuple = get_logger(__name__) class __snake_case : _a = '''dummy_data''' _a = '''datasets''' _a = False def __init__( self : Optional[Any] , A_ : str , A_ : str , A_ : Union[Version, str] , A_ : Optional[str] = None , A_ : bool = False , A_ : bool = True , A_ : Optional[List[Callable]] = None , ): lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Any = dataset_name lowerCAmelCase_ : Union[str, Any] = cache_dir lowerCAmelCase_ : List[Any] = use_local_dummy_data lowerCAmelCase_ : Optional[Any] = config # download_callbacks take a single url as input lowerCAmelCase_ : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root lowerCAmelCase_ : Tuple = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general lowerCAmelCase_ : int = str(A_) # to be downloaded lowerCAmelCase_ : Dict = None lowerCAmelCase_ : Optional[int] = None @property def UpperCAmelCase__ ( self : List[str]): if self._dummy_file is None: lowerCAmelCase_ : int = self.download_dummy_data() return self._dummy_file @property def UpperCAmelCase__ ( self : str): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('''dummy''' , self.config.name , self.version_name) # structure is dummy / version_name return os.path.join('''dummy''' , self.version_name) @property def UpperCAmelCase__ ( self : str): return os.path.join(self.dummy_data_folder , '''dummy_data.zip''') def UpperCAmelCase__ ( self : Any): lowerCAmelCase_ : Any = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) lowerCAmelCase_ : Union[str, Any] = cached_path( A_ , cache_dir=self.cache_dir , extract_compressed_file=A_ , force_extract=A_) return os.path.join(A_ , self.dummy_file_name) @property def UpperCAmelCase__ ( self : List[str]): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file) @property def UpperCAmelCase__ ( self : Optional[int]): if self._bucket_url is None: lowerCAmelCase_ : str = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '''/''')) return self._bucket_url @property def UpperCAmelCase__ ( self : List[Any]): # return full path if its a dir if os.path.isdir(self.dummy_file): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '''/''').split('''/''')[:-1]) def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Dict , *A_ : List[Any]): if self.load_existing_dummy_data: # dummy data is downloaded and tested lowerCAmelCase_ : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned lowerCAmelCase_ : Optional[int] = self.dummy_file_name # special case when data_url is a dict if isinstance(A_ , A_): return self.create_dummy_data_dict(A_ , A_) elif isinstance(A_ , (list, tuple)): return self.create_dummy_data_list(A_ , A_) else: return self.create_dummy_data_single(A_ , A_) def UpperCAmelCase__ ( self : Optional[int] , A_ : Tuple , *A_ : int): return self.download_and_extract(A_) def UpperCAmelCase__ ( self : Tuple , A_ : List[str] , A_ : Optional[Any]): return self.download_and_extract(A_) def UpperCAmelCase__ ( self : int , A_ : Optional[int] , *A_ : str , **A_ : List[Any]): return path def UpperCAmelCase__ ( self : Tuple): return {} def UpperCAmelCase__ ( self : Optional[Any] , A_ : Union[str, Any] , A_ : List[Any]): lowerCAmelCase_ : Union[str, Any] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(A_ , A_): for single_url in single_urls: download_callback(A_) else: lowerCAmelCase_ : Any = single_urls download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(A_ , A_): lowerCAmelCase_ : Any = [os.path.join(A_ , urllib.parse.quote_plus(Path(A_).name)) for x in single_urls] else: lowerCAmelCase_ : Optional[int] = single_urls lowerCAmelCase_ : List[str] = os.path.join(A_ , urllib.parse.quote_plus(Path(A_).name)) lowerCAmelCase_ : Dict = value # make sure that values are unique if all(isinstance(A_ , A_) for i in dummy_data_dict.values()) and len(set(dummy_data_dict.values())) < len( dummy_data_dict.values()): # append key to value to make its name unique lowerCAmelCase_ : Tuple = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def UpperCAmelCase__ ( self : Dict , A_ : List[str] , A_ : str): lowerCAmelCase_ : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one lowerCAmelCase_ : str = all(bool(re.findall('''[0-9]{3,}-of-[0-9]{3,}''' , A_)) for url in data_url) lowerCAmelCase_ : Optional[Any] = all( url.startswith('''https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed''') for url in data_url) if data_url and (is_tf_records or is_pubmed_records): lowerCAmelCase_ : Any = [data_url[0]] * len(A_) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowerCAmelCase_ : int = os.path.join(A_ , urllib.parse.quote_plus(single_url.split('''/''')[-1])) dummy_data_list.append(A_) return dummy_data_list def UpperCAmelCase__ ( self : List[str] , A_ : Optional[Any] , A_ : Tuple): for download_callback in self.download_callbacks: download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowerCAmelCase_ : Tuple = os.path.join(A_ , urllib.parse.quote_plus(data_url.split('''/''')[-1])) if os.path.exists(A_) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def UpperCAmelCase__ ( self : int): pass def UpperCAmelCase__ ( self : Optional[int]): pass def UpperCAmelCase__ ( self : List[str] , A_ : str): def _iter_archive_members(A_ : Any): # this preserves the order of the members inside the ZIP archive lowerCAmelCase_ : Optional[int] = Path(self.dummy_file).parent lowerCAmelCase_ : Optional[int] = path.relative_to(A_) with ZipFile(self.local_path_to_dummy_data) as zip_file: lowerCAmelCase_ : Tuple = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix()): yield dummy_parent_path.joinpath(A_) lowerCAmelCase_ : List[Any] = Path(A_) lowerCAmelCase_ : Optional[int] = _iter_archive_members(A_) if self.use_local_dummy_data else path.rglob('''*''') for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('''.''', '''__''')): yield file_path.relative_to(A_).as_posix(), file_path.open('''rb''') def UpperCAmelCase__ ( self : Dict , A_ : Any): if not isinstance(A_ , A_): lowerCAmelCase_ : Dict = [paths] for path in paths: if os.path.isfile(A_): if os.path.basename(A_).startswith(('''.''', '''__''')): return yield path else: for dirpath, dirnames, filenames in os.walk(A_): if os.path.basename(A_).startswith(('''.''', '''__''')): continue dirnames.sort() for filename in sorted(A_): if filename.startswith(('''.''', '''__''')): continue yield os.path.join(A_ , A_)
103
0
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> float: snake_case_ = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase ( ) -> List[Any]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
360
"""simple docstring""" import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process __UpperCamelCase = logging.getLogger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> Any: return (preds == labels).mean() @dataclass class UpperCamelCase : SCREAMING_SNAKE_CASE_ = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) SCREAMING_SNAKE_CASE_ = field( default=lowerCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE_ = field( default=lowerCAmelCase__ , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE_ = field( default=lowerCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCamelCase : SCREAMING_SNAKE_CASE_ = field(metadata={"help": "The name of the task to train on: " + ", ".join(processors.keys() )} ) SCREAMING_SNAKE_CASE_ = field(metadata={"help": "Should contain the data files for the task."} ) SCREAMING_SNAKE_CASE_ = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) SCREAMING_SNAKE_CASE_ = field( default=lowerCAmelCase__ , metadata={"help": "Overwrite the cached training and evaluation sets"} ) def UpperCAmelCase ( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. Use' ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , UpperCAmelCase ) # Set seed set_seed(training_args.seed ) try: snake_case_ = processors[data_args.task_name]() snake_case_ = processor.get_labels() snake_case_ = len(UpperCAmelCase ) except KeyError: raise ValueError('Task not found: %s' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=UpperCAmelCase , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=UpperCAmelCase , cache_dir=model_args.cache_dir , ) # Get datasets snake_case_ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case_ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(UpperCAmelCase ) -> Dict: snake_case_ = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(UpperCAmelCase , p.label_ids )} # Data collator snake_case_ = DataCollatorWithPadding(UpperCAmelCase , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case_ = Trainer( model=UpperCAmelCase , args=UpperCAmelCase , train_dataset=UpperCAmelCase , eval_dataset=UpperCAmelCase , compute_metrics=UpperCAmelCase , data_collator=UpperCAmelCase , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case_ = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case_ = trainer.evaluate() snake_case_ = os.path.join(training_args.output_dir , 'eval_results.txt' ) if trainer.is_world_master(): with open(UpperCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key, value in result.items(): logger.info(' %s = %s' , UpperCAmelCase , UpperCAmelCase ) writer.write('%s = %s\n' % (key, value) ) results.update(UpperCAmelCase ) return results def UpperCAmelCase ( UpperCAmelCase ) -> Optional[int]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
312
0
"""simple docstring""" from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : jnp.ndarray SCREAMING_SNAKE_CASE__ : jnp.ndarray class A_ (nn.Module ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int SCREAMING_SNAKE_CASE__ : Tuple[int] = (16, 32, 96, 256) SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase_ : Union[str, Any] = [] for i in range(len(self.block_out_channels ) - 1 ): UpperCAmelCase_ : int = self.block_out_channels[i] UpperCAmelCase_ : Any = self.block_out_channels[i + 1] UpperCAmelCase_ : Union[str, Any] = nn.Conv( lowercase_ , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(lowercase_ ) UpperCAmelCase_ : str = nn.Conv( lowercase_ , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(lowercase_ ) UpperCAmelCase_ : List[Any] = blocks UpperCAmelCase_ : List[Any] = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.conv_in(lowercase_ ) UpperCAmelCase_ : List[str] = nn.silu(lowercase_ ) for block in self.blocks: UpperCAmelCase_ : Optional[Any] = block(lowercase_ ) UpperCAmelCase_ : Optional[Any] = nn.silu(lowercase_ ) UpperCAmelCase_ : Tuple = self.conv_out(lowercase_ ) return embedding @flax_register_to_config class A_ (nn.Module ,lowercase__ ,lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = 32 SCREAMING_SNAKE_CASE__ : int = 4 SCREAMING_SNAKE_CASE__ : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) SCREAMING_SNAKE_CASE__ : Union[bool, Tuple[bool]] = False SCREAMING_SNAKE_CASE__ : Tuple[int] = (320, 640, 1280, 1280) SCREAMING_SNAKE_CASE__ : int = 2 SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int]] = 8 SCREAMING_SNAKE_CASE__ : Optional[Union[int, Tuple[int]]] = None SCREAMING_SNAKE_CASE__ : int = 1280 SCREAMING_SNAKE_CASE__ : float = 0.0 SCREAMING_SNAKE_CASE__ : bool = False SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa SCREAMING_SNAKE_CASE__ : bool = True SCREAMING_SNAKE_CASE__ : int = 0 SCREAMING_SNAKE_CASE__ : str = "rgb" SCREAMING_SNAKE_CASE__ : Tuple[int] = (16, 32, 96, 256) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" # init input tensors UpperCAmelCase_ : Tuple = (1, self.in_channels, self.sample_size, self.sample_size) UpperCAmelCase_ : int = jnp.zeros(lowercase_ , dtype=jnp.floataa ) UpperCAmelCase_ : Any = jnp.ones((1,) , dtype=jnp.intaa ) UpperCAmelCase_ : Dict = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) UpperCAmelCase_ : Union[str, Any] = (1, 3, self.sample_size * 8, self.sample_size * 8) UpperCAmelCase_ : Tuple = jnp.zeros(lowercase_ , dtype=jnp.floataa ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = jax.random.split(lowercase_ ) UpperCAmelCase_ : Any = {"params": params_rng, "dropout": dropout_rng} return self.init(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ )["params"] def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.block_out_channels UpperCAmelCase_ : Any = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCAmelCase_ : Any = self.num_attention_heads or self.attention_head_dim # input UpperCAmelCase_ : Tuple = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time UpperCAmelCase_ : str = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) UpperCAmelCase_ : int = FlaxTimestepEmbedding(lowercase_ , dtype=self.dtype ) UpperCAmelCase_ : Tuple = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) UpperCAmelCase_ : Optional[Any] = self.only_cross_attention if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = (only_cross_attention,) * len(self.down_block_types ) if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Any = (num_attention_heads,) * len(self.down_block_types ) # down UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Tuple = [] UpperCAmelCase_ : str = block_out_channels[0] UpperCAmelCase_ : Optional[int] = nn.Conv( lowercase_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(lowercase_ ) for i, down_block_type in enumerate(self.down_block_types ): UpperCAmelCase_ : Tuple = output_channel UpperCAmelCase_ : Tuple = block_out_channels[i] UpperCAmelCase_ : str = i == len(lowercase_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCAmelCase_ : Any = FlaxCrossAttnDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: UpperCAmelCase_ : str = FlaxDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(lowercase_ ) for _ in range(self.layers_per_block ): UpperCAmelCase_ : Tuple = nn.Conv( lowercase_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(lowercase_ ) if not is_final_block: UpperCAmelCase_ : str = nn.Conv( lowercase_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(lowercase_ ) UpperCAmelCase_ : List[Any] = down_blocks UpperCAmelCase_ : Tuple = controlnet_down_blocks # mid UpperCAmelCase_ : Any = block_out_channels[-1] UpperCAmelCase_ : Union[str, Any] = FlaxUNetMidBlockaDCrossAttn( in_channels=lowercase_ , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) UpperCAmelCase_ : List[str] = nn.Conv( lowercase_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = 1.0 , lowercase_ = True , lowercase_ = False , ): """simple docstring""" UpperCAmelCase_ : Any = self.controlnet_conditioning_channel_order if channel_order == "bgr": UpperCAmelCase_ : List[str] = jnp.flip(lowercase_ , axis=1 ) # 1. time if not isinstance(lowercase_ , jnp.ndarray ): UpperCAmelCase_ : Any = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(lowercase_ , jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCAmelCase_ : str = timesteps.astype(dtype=jnp.floataa ) UpperCAmelCase_ : Optional[int] = jnp.expand_dims(lowercase_ , 0 ) UpperCAmelCase_ : Optional[int] = self.time_proj(lowercase_ ) UpperCAmelCase_ : str = self.time_embedding(lowercase_ ) # 2. pre-process UpperCAmelCase_ : str = jnp.transpose(lowercase_ , (0, 2, 3, 1) ) UpperCAmelCase_ : Tuple = self.conv_in(lowercase_ ) UpperCAmelCase_ : int = jnp.transpose(lowercase_ , (0, 2, 3, 1) ) UpperCAmelCase_ : int = self.controlnet_cond_embedding(lowercase_ ) sample += controlnet_cond # 3. down UpperCAmelCase_ : Optional[int] = (sample,) for down_block in self.down_blocks: if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = down_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) else: UpperCAmelCase_ , UpperCAmelCase_ : str = down_block(lowercase_ , lowercase_ , deterministic=not train ) down_block_res_samples += res_samples # 4. mid UpperCAmelCase_ : Union[str, Any] = self.mid_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) # 5. contronet blocks UpperCAmelCase_ : Tuple = () for down_block_res_sample, controlnet_block in zip(lowercase_ , self.controlnet_down_blocks ): UpperCAmelCase_ : List[Any] = controlnet_block(lowercase_ ) controlnet_down_block_res_samples += (down_block_res_sample,) UpperCAmelCase_ : List[str] = controlnet_down_block_res_samples UpperCAmelCase_ : List[str] = self.controlnet_mid_block(lowercase_ ) # 6. scaling UpperCAmelCase_ : Optional[int] = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=lowercase_ , mid_block_res_sample=lowercase_ )
61
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case :Any = logging.get_logger(__name__) __snake_case :Optional[Any] = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } __snake_case :List[Any] = { '''b0''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def __snake_case ( _UpperCAmelCase ): __a = EfficientNetConfig() __a = CONFIG_MAP[model_name]['''hidden_dim'''] __a = CONFIG_MAP[model_name]['''width_coef'''] __a = CONFIG_MAP[model_name]['''depth_coef'''] __a = CONFIG_MAP[model_name]['''image_size'''] __a = CONFIG_MAP[model_name]['''dropout_rate'''] __a = CONFIG_MAP[model_name]['''dw_padding'''] __a = '''huggingface/label-files''' __a = '''imagenet-1k-id2label.json''' __a = 1000 __a = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) __a = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} __a = idalabel __a = {v: k for k, v in idalabel.items()} return config def __snake_case ( ): __a = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __a = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im def __snake_case ( _UpperCAmelCase ): __a = CONFIG_MAP[model_name]['''image_size'''] __a = EfficientNetImageProcessor( size={'''height''': size, '''width''': size} , image_mean=[0.4_85, 0.4_56, 0.4_06] , image_std=[0.47_85_39_44, 0.4_73_28_64, 0.47_43_41_63] , do_center_crop=_UpperCAmelCase , ) return preprocessor def __snake_case ( _UpperCAmelCase ): __a = [v.split('''_''' )[0].split('''block''' )[1] for v in original_param_names if v.startswith('''block''' )] __a = sorted(set(_UpperCAmelCase ) ) __a = len(_UpperCAmelCase ) __a = {b: str(_UpperCAmelCase ) for b, i in zip(_UpperCAmelCase , range(_UpperCAmelCase ) )} __a = [] rename_keys.append(('''stem_conv/kernel:0''', '''embeddings.convolution.weight''') ) rename_keys.append(('''stem_bn/gamma:0''', '''embeddings.batchnorm.weight''') ) rename_keys.append(('''stem_bn/beta:0''', '''embeddings.batchnorm.bias''') ) rename_keys.append(('''stem_bn/moving_mean:0''', '''embeddings.batchnorm.running_mean''') ) rename_keys.append(('''stem_bn/moving_variance:0''', '''embeddings.batchnorm.running_var''') ) for b in block_names: __a = block_name_mapping[b] rename_keys.append((f'block{b}_expand_conv/kernel:0', f'encoder.blocks.{hf_b}.expansion.expand_conv.weight') ) rename_keys.append((f'block{b}_expand_bn/gamma:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.weight') ) rename_keys.append((f'block{b}_expand_bn/beta:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.bias') ) rename_keys.append( (f'block{b}_expand_bn/moving_mean:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_mean') ) rename_keys.append( (f'block{b}_expand_bn/moving_variance:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_var') ) rename_keys.append( (f'block{b}_dwconv/depthwise_kernel:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight') ) rename_keys.append((f'block{b}_bn/gamma:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight') ) rename_keys.append((f'block{b}_bn/beta:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias') ) rename_keys.append( (f'block{b}_bn/moving_mean:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean') ) rename_keys.append( (f'block{b}_bn/moving_variance:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var') ) rename_keys.append((f'block{b}_se_reduce/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.weight') ) rename_keys.append((f'block{b}_se_reduce/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.bias') ) rename_keys.append((f'block{b}_se_expand/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.weight') ) rename_keys.append((f'block{b}_se_expand/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.bias') ) rename_keys.append( (f'block{b}_project_conv/kernel:0', f'encoder.blocks.{hf_b}.projection.project_conv.weight') ) rename_keys.append((f'block{b}_project_bn/gamma:0', f'encoder.blocks.{hf_b}.projection.project_bn.weight') ) rename_keys.append((f'block{b}_project_bn/beta:0', f'encoder.blocks.{hf_b}.projection.project_bn.bias') ) rename_keys.append( (f'block{b}_project_bn/moving_mean:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_mean') ) rename_keys.append( (f'block{b}_project_bn/moving_variance:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_var') ) rename_keys.append(('''top_conv/kernel:0''', '''encoder.top_conv.weight''') ) rename_keys.append(('''top_bn/gamma:0''', '''encoder.top_bn.weight''') ) rename_keys.append(('''top_bn/beta:0''', '''encoder.top_bn.bias''') ) rename_keys.append(('''top_bn/moving_mean:0''', '''encoder.top_bn.running_mean''') ) rename_keys.append(('''top_bn/moving_variance:0''', '''encoder.top_bn.running_var''') ) __a = {} for item in rename_keys: if item[0] in original_param_names: __a = '''efficientnet.''' + item[1] __a = '''classifier.weight''' __a = '''classifier.bias''' return key_mapping def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): for key, value in tf_params.items(): if "normalization" in key: continue __a = key_mapping[key] if "_conv" in key and "kernel" in key: __a = torch.from_numpy(_UpperCAmelCase ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: __a = torch.from_numpy(_UpperCAmelCase ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: __a = torch.from_numpy(np.transpose(_UpperCAmelCase ) ) else: __a = torch.from_numpy(_UpperCAmelCase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(_UpperCAmelCase ) @torch.no_grad() def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): __a = model_classes[model_name]( include_top=_UpperCAmelCase , weights='''imagenet''' , input_tensor=_UpperCAmelCase , input_shape=_UpperCAmelCase , pooling=_UpperCAmelCase , classes=1000 , classifier_activation='''softmax''' , ) __a = original_model.trainable_variables __a = original_model.non_trainable_variables __a = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: __a = param.numpy() __a = list(tf_params.keys() ) # Load HuggingFace model __a = get_efficientnet_config(_UpperCAmelCase ) __a = EfficientNetForImageClassification(_UpperCAmelCase ).eval() __a = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print('''Converting parameters...''' ) __a = rename_keys(_UpperCAmelCase ) replace_params(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Initialize preprocessor and preprocess input image __a = convert_image_processor(_UpperCAmelCase ) __a = preprocessor(images=prepare_img() , return_tensors='''pt''' ) # HF model inference hf_model.eval() with torch.no_grad(): __a = hf_model(**_UpperCAmelCase ) __a = outputs.logits.detach().numpy() # Original model inference __a = False __a = CONFIG_MAP[model_name]['''image_size'''] __a = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) __a = image.img_to_array(_UpperCAmelCase ) __a = np.expand_dims(_UpperCAmelCase , axis=0 ) __a = original_model.predict(_UpperCAmelCase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ), "The predicted logits are not the same." print('''Model outputs match!''' ) if save_model: # Create folder to save model if not os.path.isdir(_UpperCAmelCase ): os.mkdir(_UpperCAmelCase ) # Save converted model and image processor hf_model.save_pretrained(_UpperCAmelCase ) preprocessor.save_pretrained(_UpperCAmelCase ) if push_to_hub: # Push model and image processor to hub print(f'Pushing converted {model_name} to the hub...' ) __a = f'efficientnet-{model_name}' preprocessor.push_to_hub(_UpperCAmelCase ) hf_model.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": __snake_case :int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') __snake_case :Optional[int] = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
49
0
from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class __lowerCAmelCase : """simple docstring""" def __init__( self : int , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[int]=13 , lowerCAmelCase__ : int=7 , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : str=True , lowerCAmelCase__ : Tuple=99 , lowerCAmelCase__ : Optional[int]=32 , lowerCAmelCase__ : str=2 , lowerCAmelCase__ : Optional[Any]=4 , lowerCAmelCase__ : Optional[int]=37 , lowerCAmelCase__ : str="gelu" , lowerCAmelCase__ : Dict=0.1 , lowerCAmelCase__ : Any=0.1 , lowerCAmelCase__ : Dict=512 , lowerCAmelCase__ : int=16 , lowerCAmelCase__ : List[Any]=2 , lowerCAmelCase__ : Optional[Any]=0.02 , lowerCAmelCase__ : List[str]=3 , lowerCAmelCase__ : int=4 , lowerCAmelCase__ : Optional[Any]=None , ) -> List[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = 13 _UpperCamelCase = 7 _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = 99 _UpperCamelCase = 32 _UpperCamelCase = 2 _UpperCamelCase = 4 _UpperCamelCase = 37 _UpperCamelCase = '''gelu''' _UpperCamelCase = 0.1 _UpperCamelCase = 0.1 _UpperCamelCase = 512 _UpperCamelCase = 16 _UpperCamelCase = 2 _UpperCamelCase = 0.02 _UpperCamelCase = 3 _UpperCamelCase = 4 _UpperCamelCase = None def snake_case__ ( self : Dict ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _UpperCamelCase = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=lowerCAmelCase__ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Dict , lowerCAmelCase__ : str , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Any , lowerCAmelCase__ : Union[str, Any] ) -> int: '''simple docstring''' _UpperCamelCase = TFRoFormerModel(config=lowerCAmelCase__ ) _UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase = [input_ids, input_mask] _UpperCamelCase = model(lowerCAmelCase__ ) _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Union[str, Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Any , lowerCAmelCase__ : Dict , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : List[str] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = True _UpperCamelCase = TFRoFormerForCausalLM(config=lowerCAmelCase__ ) _UpperCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ )['''logits'''] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def snake_case__ ( self : str , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : int ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = TFRoFormerForMaskedLM(config=lowerCAmelCase__ ) _UpperCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : str , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : int , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Dict ) -> int: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = TFRoFormerForSequenceClassification(config=lowerCAmelCase__ ) _UpperCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self : int , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : str , lowerCAmelCase__ : Any , lowerCAmelCase__ : Union[str, Any] ) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = TFRoFormerForMultipleChoice(config=lowerCAmelCase__ ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__ ( self : str , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : str , lowerCAmelCase__ : Any , lowerCAmelCase__ : Dict , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = TFRoFormerForTokenClassification(config=lowerCAmelCase__ ) _UpperCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self : List[str] , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] ) -> List[Any]: '''simple docstring''' _UpperCamelCase = TFRoFormerForQuestionAnswering(config=lowerCAmelCase__ ) _UpperCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _UpperCamelCase = model(lowerCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__ ( self : Optional[int] ) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( __magic_name__ , __magic_name__ , unittest.TestCase ): """simple docstring""" _snake_case : List[str] = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) _snake_case : Any = ( { 'feature-extraction': TFRoFormerModel, 'fill-mask': TFRoFormerForMaskedLM, 'question-answering': TFRoFormerForQuestionAnswering, 'text-classification': TFRoFormerForSequenceClassification, 'text-generation': TFRoFormerForCausalLM, 'token-classification': TFRoFormerForTokenClassification, 'zero-shot': TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) _snake_case : Optional[int] = False _snake_case : Optional[Any] = False def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[str] ) -> str: '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def snake_case__ ( self : List[str] ) -> List[str]: '''simple docstring''' _UpperCamelCase = TFRoFormerModelTester(self ) _UpperCamelCase = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=37 ) def snake_case__ ( self : Optional[int] ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def snake_case__ ( self : str ) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def snake_case__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase__ ) def snake_case__ ( self : Tuple ) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*lowerCAmelCase__ ) def snake_case__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCAmelCase__ ) def snake_case__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase__ ) def snake_case__ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase__ ) def snake_case__ ( self : List[str] ) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase__ ) @slow def snake_case__ ( self : Optional[int] ) -> Any: '''simple docstring''' _UpperCamelCase = TFRoFormerModel.from_pretrained('''junnyu/roformer_chinese_base''' ) self.assertIsNotNone(lowerCAmelCase__ ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def snake_case__ ( self : List[str] ) -> int: '''simple docstring''' _UpperCamelCase = TFRoFormerForMaskedLM.from_pretrained('''junnyu/roformer_chinese_base''' ) _UpperCamelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) _UpperCamelCase = model(lowerCAmelCase__ )[0] # TODO Replace vocab size _UpperCamelCase = 50000 _UpperCamelCase = [1, 6, vocab_size] self.assertEqual(output.shape , lowerCAmelCase__ ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. _UpperCamelCase = tf.constant( [ [ [-0.12053341, -1.0264901, 0.29221946], [-1.5133783, 0.197433, 0.15190607], [-5.0135403, -3.900256, -0.84038764], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCAmelCase__ , atol=1e-4 ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" _snake_case : Union[str, Any] = 1e-4 def snake_case__ ( self : str ) -> List[str]: '''simple docstring''' _UpperCamelCase = tf.constant([[4, 10]] ) _UpperCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) _UpperCamelCase = emba(input_ids.shape ) _UpperCamelCase = tf.constant( [[0.0000, 0.0000, 0.0000, 1.0000, 1.0000, 1.0000], [0.8415, 0.0464, 0.0022, 0.5403, 0.9989, 1.0000]] ) tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , atol=self.tolerance ) def snake_case__ ( self : Optional[int] ) -> List[str]: '''simple docstring''' _UpperCamelCase = tf.constant( [ [0.0000, 0.0000, 0.0000, 0.0000, 0.0000], [0.8415, 0.8219, 0.8020, 0.7819, 0.7617], [0.9093, 0.9364, 0.9581, 0.9749, 0.9870], ] ) _UpperCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=512 , embedding_dim=512 ) emba([2, 16, 512] ) _UpperCamelCase = emba.weight[:3, :5] tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , atol=self.tolerance ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" _snake_case : Any = 1e-4 def snake_case__ ( self : List[str] ) -> str: '''simple docstring''' _UpperCamelCase = tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 _UpperCamelCase = -tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 _UpperCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32 , embedding_dim=64 ) _UpperCamelCase = embed_positions([2, 16, 768] )[None, None, :, :] _UpperCamelCase , _UpperCamelCase = TFRoFormerSelfAttention.apply_rotary_position_embeddings( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = tf.constant( [ [0.0000, 0.0100, 0.0200, 0.0300, 0.0400, 0.0500, 0.0600, 0.0700], [-0.2012, 0.8897, 0.0263, 0.9401, 0.2074, 0.9463, 0.3481, 0.9343], [-1.7057, 0.6271, -1.2145, 1.3897, -0.6303, 1.7647, -0.1173, 1.8985], [-2.1731, -1.6397, -2.7358, 0.2854, -2.1840, 1.7183, -1.3018, 2.4871], [0.2717, -3.6173, -2.9206, -2.1988, -3.6638, 0.3858, -2.9155, 2.2980], [3.9859, -2.1580, -0.7984, -4.4904, -4.1181, -2.0252, -4.4782, 1.1253], ] ) _UpperCamelCase = tf.constant( [ [0.0000, -0.0100, -0.0200, -0.0300, -0.0400, -0.0500, -0.0600, -0.0700], [0.2012, -0.8897, -0.0263, -0.9401, -0.2074, -0.9463, -0.3481, -0.9343], [1.7057, -0.6271, 1.2145, -1.3897, 0.6303, -1.7647, 0.1173, -1.8985], [2.1731, 1.6397, 2.7358, -0.2854, 2.1840, -1.7183, 1.3018, -2.4871], [-0.2717, 3.6173, 2.9206, 2.1988, 3.6638, -0.3858, 2.9155, -2.2980], [-3.9859, 2.1580, 0.7984, 4.4904, 4.1181, 2.0252, 4.4782, -1.1253], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , lowerCAmelCase__ , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , lowerCAmelCase__ , atol=self.tolerance )
363
'''simple docstring''' import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase__ : Optional[Any] = get_tests_dir('fixtures/test_sentencepiece.model') lowercase__ : Any = get_tests_dir('fixtures/test_sentencepiece_bpe.model') lowercase__ : Tuple = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): """simple docstring""" _snake_case : Union[str, Any] = CamembertTokenizer _snake_case : str = CamembertTokenizerFast _snake_case : int = True _snake_case : List[str] = True def snake_case__ ( self : Dict ) -> Any: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _UpperCamelCase = CamembertTokenizer(lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' _UpperCamelCase = '''<pad>''' _UpperCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def snake_case__ ( self : Dict ) -> List[Any]: '''simple docstring''' _UpperCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>NOTUSED''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCAmelCase__ ) , 1004 ) def snake_case__ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1005 ) def snake_case__ ( self : int ) -> Tuple: '''simple docstring''' _UpperCamelCase = CamembertTokenizer(lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) _UpperCamelCase = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) _UpperCamelCase = '''I was born in 92000, and this is falsé.''' _UpperCamelCase = tokenizer.encode(lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.encode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) _UpperCamelCase = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def snake_case__ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' if not self.test_rust_tokenizer: return _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = '''I was born in 92000, and this is falsé.''' _UpperCamelCase = tokenizer.tokenize(lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = tokenizer.encode(lowerCAmelCase__ ) _UpperCamelCase = rust_tokenizer.encode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) @slow def snake_case__ ( self : Any ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = {'''input_ids''': [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 27575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 22804, 18818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 10326, 24, 2267, 20, 416, 5072, 15612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. _UpperCamelCase = [ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name='''camembert-base''' , revision='''3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf''' , sequences=lowerCAmelCase__ , )
287
0
'''simple docstring''' import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class __lowerCAmelCase ( _lowercase , unittest.TestCase ): """simple docstring""" _snake_case : Any = MobileBertTokenizer _snake_case : Tuple = MobileBertTokenizerFast _snake_case : List[str] = True _snake_case : int = True _snake_case : Tuple = filter_non_english _snake_case : List[Any] = 'google/mobilebert-uncased' def snake_case__ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' super().setUp() _UpperCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] _UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) _UpperCamelCase = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def snake_case__ ( self : int , lowerCAmelCase__ : int ) -> Any: '''simple docstring''' _UpperCamelCase = '''UNwant\u00E9d,running''' _UpperCamelCase = '''unwanted, running''' return input_text, output_text def snake_case__ ( self : Union[str, Any] ) -> int: '''simple docstring''' _UpperCamelCase = self.tokenizer_class(self.vocab_file ) _UpperCamelCase = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(__UpperCamelCase , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCamelCase ) , [9, 6, 7, 12, 10, 11] ) def snake_case__ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' if not self.test_rust_tokenizer: return _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = '''UNwant\u00E9d,running''' _UpperCamelCase = tokenizer.tokenize(__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = tokenizer.encode(__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.encode(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) # With lower casing _UpperCamelCase = self.get_tokenizer(do_lower_case=__UpperCamelCase ) _UpperCamelCase = self.get_rust_tokenizer(do_lower_case=__UpperCamelCase ) _UpperCamelCase = '''UNwant\u00E9d,running''' _UpperCamelCase = tokenizer.tokenize(__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = tokenizer.encode(__UpperCamelCase ) _UpperCamelCase = rust_tokenizer.encode(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) def snake_case__ ( self : str ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def snake_case__ ( self : Any ) -> Dict: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def snake_case__ ( self : Tuple ) -> Tuple: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase , strip_accents=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def snake_case__ ( self : Optional[Any] ) -> Dict: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase , strip_accents=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def snake_case__ ( self : Any ) -> List[Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def snake_case__ ( self : Any ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def snake_case__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase , strip_accents=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def snake_case__ ( self : int ) -> Tuple: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase , strip_accents=__UpperCamelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def snake_case__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__UpperCamelCase , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def snake_case__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] _UpperCamelCase = {} for i, token in enumerate(__UpperCamelCase ): _UpperCamelCase = i _UpperCamelCase = WordpieceTokenizer(vocab=__UpperCamelCase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) def snake_case__ ( self : Dict ) -> List[Any]: '''simple docstring''' self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def snake_case__ ( self : List[Any] ) -> int: '''simple docstring''' self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def snake_case__ ( self : Any ) -> Tuple: '''simple docstring''' self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) def snake_case__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__UpperCamelCase ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) self.assertListEqual( [rust_tokenizer.tokenize(__UpperCamelCase ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) @slow def snake_case__ ( self : Union[str, Any] ) -> Any: '''simple docstring''' _UpperCamelCase = self.tokenizer_class.from_pretrained('''google/mobilebert-uncased''' ) _UpperCamelCase = tokenizer.encode('''sequence builders''' , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__UpperCamelCase ) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__UpperCamelCase , __UpperCamelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def snake_case__ ( self : Tuple ) -> Dict: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" _UpperCamelCase = tokenizer_r.encode_plus( __UpperCamelCase , return_attention_mask=__UpperCamelCase , return_token_type_ids=__UpperCamelCase , return_offsets_mapping=__UpperCamelCase , add_special_tokens=__UpperCamelCase , ) _UpperCamelCase = tokenizer_r.do_lower_case if hasattr(__UpperCamelCase , '''do_lower_case''' ) else False _UpperCamelCase = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''A'''), ((1, 2), ''','''), ((3, 5), '''na'''), ((5, 6), '''##ï'''), ((6, 8), '''##ve'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''Allen'''), ((21, 23), '''##NL'''), ((23, 24), '''##P'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''a'''), ((1, 2), ''','''), ((3, 8), '''naive'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''allen'''), ((21, 23), '''##nl'''), ((23, 24), '''##p'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['''offset_mapping'''] ) def snake_case__ ( self : Any ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ['''的''', '''人''', '''有'''] _UpperCamelCase = ''''''.join(__UpperCamelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCamelCase = True _UpperCamelCase = self.tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = tokenizer_p.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer_r.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer_r.convert_ids_to_tokens(__UpperCamelCase ) _UpperCamelCase = tokenizer_p.convert_ids_to_tokens(__UpperCamelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = False _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = self.tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) _UpperCamelCase = tokenizer_r.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer_p.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) _UpperCamelCase = tokenizer_r.convert_ids_to_tokens(__UpperCamelCase ) _UpperCamelCase = tokenizer_p.convert_ids_to_tokens(__UpperCamelCase ) # it is expected that only the first Chinese character is not preceded by "##". _UpperCamelCase = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(__UpperCamelCase ) ] self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase )
324
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCAmelCase = """base_with_context""" def lowercase ( a__ : Optional[Any] , a__ : Optional[int] ) -> int: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''token_embedder''']['''embedding'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) for lyr_num, lyr in enumerate(model.encoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = ly_weight['''attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def lowercase ( a__ : List[Any] , a__ : Dict ) -> Optional[Any]: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''input_proj''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) for lyr_num, lyr in enumerate(model.encoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = ly_weight['''attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def lowercase ( a__ : List[Any] , a__ : Union[str, Any] ) -> str: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''continuous_inputs_projection''']['''kernel'''].T ) ) for lyr_num, lyr in enumerate(model.decoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_self_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_0''']['''DenseGeneral_0''']['''kernel'''].T ) ) _UpperCamelCase = ly_weight['''self_attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = ly_weight['''MultiHeadDotProductAttention_0'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_cross_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_1''']['''DenseGeneral_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''decoder_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''spec_out_dense''']['''kernel'''].T ) ) return model def lowercase ( a__ : Union[str, Any] ) -> int: _UpperCamelCase = checkpoints.load_tax_checkpoint(args.checkpoint_path ) _UpperCamelCase = jnp.tree_util.tree_map(onp.array , a__ ) _UpperCamelCase = [ '''from __gin__ import dynamic_registration''', '''from music_spectrogram_diffusion.models.diffusion import diffusion_utils''', '''diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0''', '''diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()''', ] _UpperCamelCase = os.path.join(args.checkpoint_path , '''..''' , '''config.gin''' ) _UpperCamelCase = inference.parse_training_gin_file(a__ , a__ ) _UpperCamelCase = inference.InferenceModel(args.checkpoint_path , a__ ) _UpperCamelCase = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' , variance_type='''fixed_large''' ) _UpperCamelCase = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['''inputs'''] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) _UpperCamelCase = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['''targets_context'''] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) _UpperCamelCase = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['''targets_context'''] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) _UpperCamelCase = load_notes_encoder(ta_checkpoint['''target''']['''token_encoder'''] , a__ ) _UpperCamelCase = load_continuous_encoder(ta_checkpoint['''target''']['''continuous_encoder'''] , a__ ) _UpperCamelCase = load_decoder(ta_checkpoint['''target''']['''decoder'''] , a__ ) _UpperCamelCase = OnnxRuntimeModel.from_pretrained('''kashif/soundstream_mel_decoder''' ) _UpperCamelCase = SpectrogramDiffusionPipeline( notes_encoder=a__ , continuous_encoder=a__ , decoder=a__ , scheduler=a__ , melgan=a__ , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=F'''{MODEL}/checkpoint_500000''', type=str, required=False, help="""Path to the original jax model checkpoint.""", ) UpperCAmelCase = parser.parse_args() main(args)
256
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def __UpperCamelCase( lowerCAmelCase__ : str , lowerCAmelCase__ : str , lowerCAmelCase__ : str , lowerCAmelCase__ : PreTrainedTokenizer , lowerCAmelCase__ : int , lowerCAmelCase__ : Optional[int] = None , ): __a : Optional[int] = {} if train_file is not None: __a : Dict = [train_file] if eval_file is not None: __a : int = [eval_file] if test_file is not None: __a : Any = [test_file] __a : Tuple = datasets.load_dataset('''csv''' , data_files=lowerCAmelCase__ ) __a : List[Any] = list(ds[list(files.keys() )[0]].features.keys() ) __a : Any = features_name.pop(lowerCAmelCase__ ) __a : int = list(set(ds[list(files.keys() )[0]][label_name] ) ) __a : str = {label: i for i, label in enumerate(lowerCAmelCase__ )} __a : Tuple = tokenizer.model_input_names __a : Optional[Any] = {} if len(lowerCAmelCase__ ) == 1: for k in files.keys(): __a : Optional[Any] = ds[k].map( lambda lowerCAmelCase__ : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding='''max_length''' ) , batched=lowerCAmelCase__ , ) elif len(lowerCAmelCase__ ) == 2: for k in files.keys(): __a : str = ds[k].map( lambda lowerCAmelCase__ : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding='''max_length''' , ) , batched=lowerCAmelCase__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: __a : List[str] = {k: v for k, v in ex.items() if k in input_names} __a : List[Any] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: __a : Dict = {k: v for k, v in ex.items() if k in input_names} __a : List[Any] = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: __a : List[Any] = {k: v for k, v in ex.items() if k in input_names} __a : Dict = labelaid[ex[label_name]] yield (d, label) __a : List[Any] = ( tf.data.Dataset.from_generator( lowerCAmelCase__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: __a : Tuple = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) __a : Tuple = ( tf.data.Dataset.from_generator( lowerCAmelCase__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: __a : Tuple = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) __a : int = ( tf.data.Dataset.from_generator( lowerCAmelCase__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: __a : Union[str, Any] = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid lowercase__ =logging.getLogger(__name__) @dataclass class UpperCamelCase__ : _SCREAMING_SNAKE_CASE : int = field(metadata={"help": "Which column contains the label"} ) _SCREAMING_SNAKE_CASE : str = field(default=__lowercase ,metadata={"help": "The path of the training file"} ) _SCREAMING_SNAKE_CASE : Optional[str] = field(default=__lowercase ,metadata={"help": "The path of the development file"} ) _SCREAMING_SNAKE_CASE : Optional[str] = field(default=__lowercase ,metadata={"help": "The path of the test file"} ) _SCREAMING_SNAKE_CASE : int = field( default=128 ,metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } ,) _SCREAMING_SNAKE_CASE : bool = field( default=__lowercase ,metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class UpperCamelCase__ : _SCREAMING_SNAKE_CASE : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _SCREAMING_SNAKE_CASE : Optional[str] = field( default=__lowercase ,metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _SCREAMING_SNAKE_CASE : Optional[str] = field( default=__lowercase ,metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) _SCREAMING_SNAKE_CASE : bool = field(default=__lowercase ,metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. _SCREAMING_SNAKE_CASE : Optional[str] = field( default=__lowercase ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} ,) def __UpperCamelCase( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __a : Dict = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) __a : Dict = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info( f"n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, " f"16-bits training: {training_args.fpaa}" ) logger.info(f"Training/evaluation parameters {training_args}" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __a : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) __a : List[str] = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=lowerCAmelCase__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) __a : List[str] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(lowerCAmelCase__ ) , labelaid=lowerCAmelCase__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='''text-classification''' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): __a : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('''.bin''' in model_args.model_name_or_path ) , config=lowerCAmelCase__ , cache_dir=model_args.cache_dir , ) def compute_metrics(lowerCAmelCase__ : EvalPrediction ) -> Dict: __a : int = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer __a : Optional[Any] = TFTrainer( model=lowerCAmelCase__ , args=lowerCAmelCase__ , train_dataset=lowerCAmelCase__ , eval_dataset=lowerCAmelCase__ , compute_metrics=lowerCAmelCase__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation __a : str = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) __a : str = trainer.evaluate() __a : int = os.path.join(training_args.output_dir , '''eval_results.txt''' ) with open(lowerCAmelCase__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(f" {key} = {value}" ) writer.write(f"{key} = {value}\n" ) results.update(lowerCAmelCase__ ) return results if __name__ == "__main__": main()
352
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) lowercase__ ={ 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =['LayoutLMv2FeatureExtractor'] lowercase__ =['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys lowercase__ =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
90
0
import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class __lowercase ( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Optional[int] = FlaxAutoencoderKL @property def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Tuple = 4 SCREAMING_SNAKE_CASE_: List[str] = 3 SCREAMING_SNAKE_CASE_: str = (32, 32) SCREAMING_SNAKE_CASE_: Dict = jax.random.PRNGKey(0) SCREAMING_SNAKE_CASE_: List[str] = jax.random.uniform(lowerCAmelCase__ , ((batch_size, num_channels) + sizes)) return {"sample": image, "prng_key": prng_key} def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: Dict = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } SCREAMING_SNAKE_CASE_: Any = self.dummy_input return init_dict, inputs_dict
13
"""simple docstring""" import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowercase__ : Optional[Any] = logging.get_logger(__name__) class UpperCamelCase__ ( lowercase_ ): """simple docstring""" _SCREAMING_SNAKE_CASE = """linear""" _SCREAMING_SNAKE_CASE = """cosine""" _SCREAMING_SNAKE_CASE = """cosine_with_restarts""" _SCREAMING_SNAKE_CASE = """polynomial""" _SCREAMING_SNAKE_CASE = """constant""" _SCREAMING_SNAKE_CASE = """constant_with_warmup""" _SCREAMING_SNAKE_CASE = """piecewise_constant""" def UpperCamelCase_ ( lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : int = -1 ) -> Tuple: """simple docstring""" return LambdaLR(lowerCAmelCase__ , lambda lowerCAmelCase__ : 1 , last_epoch=lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : int , lowerCAmelCase__ : int = -1 ) -> str: """simple docstring""" def lr_lambda(lowerCAmelCase__ : int ): if current_step < num_warmup_steps: return float(lowerCAmelCase__ ) / float(max(1.0 , lowerCAmelCase__ ) ) return 1.0 return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , last_epoch=lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : str , lowerCAmelCase__ : int = -1 ) -> int: """simple docstring""" lowerCAmelCase_ : str = {} lowerCAmelCase_ : Optional[Any] = step_rules.split(',' ) for rule_str in rule_list[:-1]: lowerCAmelCase_ ,lowerCAmelCase_ : Optional[int] = rule_str.split(':' ) lowerCAmelCase_ : List[str] = int(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = float(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = value lowerCAmelCase_ : str = float(rule_list[-1] ) def create_rules_function(lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : int ): def rule_func(lowerCAmelCase__ : int ) -> float: lowerCAmelCase_ : str = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(lowerCAmelCase__ ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func lowerCAmelCase_ : Dict = create_rules_function(lowerCAmelCase__ , lowerCAmelCase__ ) return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , last_epoch=lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Union[str, Any]=-1 ) -> Any: """simple docstring""" def lr_lambda(lowerCAmelCase__ : int ): if current_step < num_warmup_steps: return float(lowerCAmelCase__ ) / float(max(1 , lowerCAmelCase__ ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : int , lowerCAmelCase__ : int , lowerCAmelCase__ : float = 0.5 , lowerCAmelCase__ : int = -1 ) -> Union[str, Any]: """simple docstring""" def lr_lambda(lowerCAmelCase__ : Any ): if current_step < num_warmup_steps: return float(lowerCAmelCase__ ) / float(max(1 , lowerCAmelCase__ ) ) lowerCAmelCase_ : List[Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(lowerCAmelCase__ ) * 2.0 * progress )) ) return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : int , lowerCAmelCase__ : int , lowerCAmelCase__ : int = 1 , lowerCAmelCase__ : int = -1 ) -> int: """simple docstring""" def lr_lambda(lowerCAmelCase__ : str ): if current_step < num_warmup_steps: return float(lowerCAmelCase__ ) / float(max(1 , lowerCAmelCase__ ) ) lowerCAmelCase_ : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(lowerCAmelCase__ ) * progress) % 1.0) )) ) return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Tuple=1e-7 , lowerCAmelCase__ : Union[str, Any]=1.0 , lowerCAmelCase__ : int=-1 ) -> Optional[Any]: """simple docstring""" lowerCAmelCase_ : Any = optimizer.defaults['lr'] if not (lr_init > lr_end): raise ValueError(f"lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})" ) def lr_lambda(lowerCAmelCase__ : int ): if current_step < num_warmup_steps: return float(lowerCAmelCase__ ) / float(max(1 , lowerCAmelCase__ ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: lowerCAmelCase_ : Any = lr_init - lr_end lowerCAmelCase_ : int = num_training_steps - num_warmup_steps lowerCAmelCase_ : Dict = 1 - (current_step - num_warmup_steps) / decay_steps lowerCAmelCase_ : Dict = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__ : Any = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def UpperCamelCase_ ( lowerCAmelCase__ : Union[str, SchedulerType] , lowerCAmelCase__ : Optimizer , lowerCAmelCase__ : Optional[str] = None , lowerCAmelCase__ : Optional[int] = None , lowerCAmelCase__ : Optional[int] = None , lowerCAmelCase__ : int = 1 , lowerCAmelCase__ : float = 1.0 , lowerCAmelCase__ : int = -1 , ) -> int: """simple docstring""" lowerCAmelCase_ : List[str] = SchedulerType(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(lowerCAmelCase__ , last_epoch=lowerCAmelCase__ ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(lowerCAmelCase__ , step_rules=lowerCAmelCase__ , last_epoch=lowerCAmelCase__ ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"{name} requires `num_warmup_steps`, please provide that argument." ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(lowerCAmelCase__ , num_warmup_steps=lowerCAmelCase__ , last_epoch=lowerCAmelCase__ ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"{name} requires `num_training_steps`, please provide that argument." ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( lowerCAmelCase__ , num_warmup_steps=lowerCAmelCase__ , num_training_steps=lowerCAmelCase__ , num_cycles=lowerCAmelCase__ , last_epoch=lowerCAmelCase__ , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( lowerCAmelCase__ , num_warmup_steps=lowerCAmelCase__ , num_training_steps=lowerCAmelCase__ , power=lowerCAmelCase__ , last_epoch=lowerCAmelCase__ , ) return schedule_func( lowerCAmelCase__ , num_warmup_steps=lowerCAmelCase__ , num_training_steps=lowerCAmelCase__ , last_epoch=lowerCAmelCase__ )
224
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType lowercase_ = logging.get_logger(__name__) lowercase_ = { """openai/imagegpt-small""": """""", """openai/imagegpt-medium""": """""", """openai/imagegpt-large""": """""", } class _snake_case ( lowercase__): UpperCamelCase__ : List[Any] ="""imagegpt""" UpperCamelCase__ : List[str] =["""past_key_values"""] UpperCamelCase__ : str ={ """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Union[str, Any], __lowercase : List[str]=512 + 1, __lowercase : Optional[Any]=32 * 32, __lowercase : str=512, __lowercase : List[str]=24, __lowercase : int=8, __lowercase : List[Any]=None, __lowercase : int="quick_gelu", __lowercase : Dict=0.1, __lowercase : Dict=0.1, __lowercase : Optional[Any]=0.1, __lowercase : Union[str, Any]=1e-5, __lowercase : Any=0.02, __lowercase : Union[str, Any]=True, __lowercase : Dict=True, __lowercase : int=False, __lowercase : int=False, __lowercase : Any=False, **__lowercase : Tuple, ): lowercase__ = vocab_size lowercase__ = n_positions lowercase__ = n_embd lowercase__ = n_layer lowercase__ = n_head lowercase__ = n_inner lowercase__ = activation_function lowercase__ = resid_pdrop lowercase__ = embd_pdrop lowercase__ = attn_pdrop lowercase__ = layer_norm_epsilon lowercase__ = initializer_range lowercase__ = scale_attn_weights lowercase__ = use_cache lowercase__ = scale_attn_by_inverse_layer_idx lowercase__ = reorder_and_upcast_attn lowercase__ = tie_word_embeddings super().__init__(tie_word_embeddings=__lowercase, **__lowercase ) class _snake_case ( lowercase__): @property def A__ ( self : Union[str, Any] ): return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ] ) def A__ ( self : Tuple, __lowercase : "FeatureExtractionMixin", __lowercase : int = 1, __lowercase : int = -1, __lowercase : bool = False, __lowercase : Optional["TensorType"] = None, __lowercase : int = 3, __lowercase : int = 32, __lowercase : int = 32, ): lowercase__ = self._generate_dummy_images(__lowercase, __lowercase, __lowercase, __lowercase ) lowercase__ = dict(preprocessor(images=__lowercase, return_tensors=__lowercase ) ) return inputs
224
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class _snake_case ( lowercase__): UpperCamelCase__ : Optional[Any] ="""transfo-xl""" UpperCamelCase__ : Dict =["""mems"""] UpperCamelCase__ : Optional[int] ={ """n_token""": """vocab_size""", """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Optional[Any], __lowercase : Optional[Any]=26_7735, __lowercase : int=[2_0000, 4_0000, 20_0000], __lowercase : Union[str, Any]=1024, __lowercase : Tuple=1024, __lowercase : Tuple=16, __lowercase : Optional[Any]=64, __lowercase : str=4096, __lowercase : Optional[int]=4, __lowercase : Union[str, Any]=False, __lowercase : Union[str, Any]=18, __lowercase : List[str]=1600, __lowercase : List[Any]=1000, __lowercase : Union[str, Any]=True, __lowercase : Tuple=True, __lowercase : Optional[Any]=0, __lowercase : List[str]=-1, __lowercase : int=True, __lowercase : Dict=0.1, __lowercase : Union[str, Any]=0.0, __lowercase : str=True, __lowercase : Optional[Any]="normal", __lowercase : str=0.01, __lowercase : Tuple=0.01, __lowercase : List[Any]=0.02, __lowercase : Any=1e-5, __lowercase : Union[str, Any]=0, **__lowercase : Union[str, Any], ): lowercase__ = vocab_size lowercase__ = [] self.cutoffs.extend(__lowercase ) if proj_share_all_but_first: lowercase__ = [False] + [True] * len(self.cutoffs ) else: lowercase__ = [False] + [False] * len(self.cutoffs ) lowercase__ = d_model lowercase__ = d_embed lowercase__ = d_head lowercase__ = d_inner lowercase__ = div_val lowercase__ = pre_lnorm lowercase__ = n_layer lowercase__ = n_head lowercase__ = mem_len lowercase__ = same_length lowercase__ = attn_type lowercase__ = clamp_len lowercase__ = sample_softmax lowercase__ = adaptive lowercase__ = dropout lowercase__ = dropatt lowercase__ = untie_r lowercase__ = init lowercase__ = init_range lowercase__ = proj_init_std lowercase__ = init_std lowercase__ = layer_norm_epsilon super().__init__(eos_token_id=__lowercase, **__lowercase ) @property def A__ ( self : Optional[Any] ): # Message copied from Transformer-XL documentation logger.info(F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def A__ ( self : List[str], __lowercase : Union[str, Any] ): # Message copied from Transformer-XL documentation raise NotImplementedError( F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
224
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ =logging.get_logger(__name__) lowercase__ ={ 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class UpperCamelCase__ ( __lowercase ): _SCREAMING_SNAKE_CASE : List[str] = "speech_to_text_2" _SCREAMING_SNAKE_CASE : List[Any] = ["past_key_values"] _SCREAMING_SNAKE_CASE : Tuple = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__(self : str , snake_case_ : Tuple=1_0_0_0_0 , snake_case_ : Union[str, Any]=6 , snake_case_ : Optional[Any]=2_0_4_8 , snake_case_ : int=4 , snake_case_ : Optional[int]=0.0 , snake_case_ : Dict=True , snake_case_ : str="relu" , snake_case_ : int=2_5_6 , snake_case_ : List[str]=0.1 , snake_case_ : List[str]=0.0 , snake_case_ : List[Any]=0.0 , snake_case_ : List[Any]=0.02 , snake_case_ : Any=2 , snake_case_ : List[str]=True , snake_case_ : Tuple=1 , snake_case_ : List[Any]=0 , snake_case_ : List[str]=2 , snake_case_ : Optional[Any]=1_0_2_4 , **snake_case_ : List[str] , ): __a : Optional[Any] = vocab_size __a : Optional[Any] = d_model __a : List[Any] = decoder_ffn_dim __a : List[Any] = decoder_layers __a : Tuple = decoder_attention_heads __a : List[Any] = dropout __a : str = attention_dropout __a : Dict = activation_dropout __a : str = activation_function __a : Tuple = init_std __a : str = decoder_layerdrop __a : Any = use_cache __a : Dict = decoder_layers __a : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __a : int = max_target_positions super().__init__( pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , **snake_case_ , )
216
import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertModel, ) class UpperCamelCase__ : def __init__(self : List[Any] , snake_case_ : int , snake_case_ : List[str]=1_3 , snake_case_ : Tuple=7 , snake_case_ : List[Any]=True , snake_case_ : List[Any]=True , snake_case_ : Dict=True , snake_case_ : Optional[int]=True , snake_case_ : str=9_9 , snake_case_ : Dict=6_4 , snake_case_ : Any=3_2 , snake_case_ : str=5 , snake_case_ : int=4 , snake_case_ : List[Any]=3_7 , snake_case_ : Any="gelu" , snake_case_ : Dict=0.1 , snake_case_ : List[str]=0.1 , snake_case_ : str=5_1_2 , snake_case_ : Any=1_6 , snake_case_ : str=2 , snake_case_ : int=0.02 , snake_case_ : Union[str, Any]=3 , snake_case_ : Optional[Any]=4 , snake_case_ : List[Any]=None , ): __a : Any = parent __a : Optional[int] = batch_size __a : Any = seq_length __a : int = is_training __a : Optional[int] = use_input_mask __a : List[Any] = use_token_type_ids __a : Dict = use_labels __a : Tuple = vocab_size __a : str = hidden_size __a : List[Any] = embedding_size __a : List[Any] = num_hidden_layers __a : str = num_attention_heads __a : str = intermediate_size __a : Union[str, Any] = hidden_act __a : Optional[Any] = hidden_dropout_prob __a : Tuple = attention_probs_dropout_prob __a : Union[str, Any] = max_position_embeddings __a : Any = type_vocab_size __a : int = type_sequence_label_size __a : int = initializer_range __a : int = num_labels __a : Union[str, Any] = num_choices __a : Dict = scope def lowerCAmelCase (self : str ): __a : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : List[Any] = None if self.use_input_mask: __a : Dict = random_attention_mask([self.batch_size, self.seq_length] ) __a : Optional[Any] = None if self.use_token_type_ids: __a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __a : Dict = None __a : List[str] = None __a : Optional[Any] = None if self.use_labels: __a : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) __a : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase (self : int ): return MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) def lowerCAmelCase (self : str , snake_case_ : Tuple , snake_case_ : List[str] , snake_case_ : int , snake_case_ : int , snake_case_ : Any , snake_case_ : List[Any] , snake_case_ : Any ): __a : Any = MobileBertModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : List[str] = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ ) __a : Optional[Any] = model(snake_case_ , token_type_ids=snake_case_ ) __a : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCAmelCase (self : Any , snake_case_ : Dict , snake_case_ : Optional[Any] , snake_case_ : Tuple , snake_case_ : Dict , snake_case_ : Tuple , snake_case_ : str , snake_case_ : List[Any] ): __a : str = MobileBertForMaskedLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : Tuple = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase (self : Tuple , snake_case_ : Any , snake_case_ : Dict , snake_case_ : List[Any] , snake_case_ : Any , snake_case_ : Union[str, Any] , snake_case_ : Optional[int] , snake_case_ : Dict ): __a : Optional[Any] = MobileBertForNextSentencePrediction(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : int = model( snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCAmelCase (self : Any , snake_case_ : Dict , snake_case_ : Optional[int] , snake_case_ : Any , snake_case_ : Any , snake_case_ : Union[str, Any] , snake_case_ : Dict , snake_case_ : Optional[Any] ): __a : str = MobileBertForPreTraining(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : Union[str, Any] = model( snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ , next_sentence_label=snake_case_ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCAmelCase (self : Dict , snake_case_ : Tuple , snake_case_ : Any , snake_case_ : Dict , snake_case_ : int , snake_case_ : int , snake_case_ : str , snake_case_ : str ): __a : str = MobileBertForQuestionAnswering(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : Optional[Any] = model( snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , start_positions=snake_case_ , end_positions=snake_case_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase (self : Optional[int] , snake_case_ : Any , snake_case_ : List[str] , snake_case_ : List[Any] , snake_case_ : str , snake_case_ : Any , snake_case_ : Tuple , snake_case_ : Optional[int] ): __a : Any = self.num_labels __a : Union[str, Any] = MobileBertForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() __a : Tuple = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase (self : List[Any] , snake_case_ : Optional[int] , snake_case_ : str , snake_case_ : Dict , snake_case_ : List[Any] , snake_case_ : Optional[Any] , snake_case_ : Dict , snake_case_ : Optional[int] ): __a : Union[str, Any] = self.num_labels __a : str = MobileBertForTokenClassification(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : Any = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase (self : Union[str, Any] , snake_case_ : List[str] , snake_case_ : Any , snake_case_ : Optional[Any] , snake_case_ : Optional[Any] , snake_case_ : Any , snake_case_ : Dict , snake_case_ : Union[str, Any] ): __a : Union[str, Any] = self.num_choices __a : List[str] = MobileBertForMultipleChoice(config=snake_case_ ) model.to(snake_case_ ) model.eval() __a : List[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : int = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : Any = model( snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase (self : Optional[Any] ): __a : Optional[Any] = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) : int = config_and_inputs __a : Union[str, Any] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase__ ( __lowercase ,__lowercase ,unittest.TestCase ): _SCREAMING_SNAKE_CASE : str = ( ( MobileBertModel, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, ) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE : Any = ( { "feature-extraction": MobileBertModel, "fill-mask": MobileBertForMaskedLM, "question-answering": MobileBertForQuestionAnswering, "text-classification": MobileBertForSequenceClassification, "token-classification": MobileBertForTokenClassification, "zero-shot": MobileBertForSequenceClassification, } if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE : Union[str, Any] = True def lowerCAmelCase (self : str , snake_case_ : Tuple , snake_case_ : List[Any] , snake_case_ : Union[str, Any]=False ): __a : List[str] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): __a : Optional[int] = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=snake_case_ ) __a : int = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case_ ) return inputs_dict def lowerCAmelCase (self : Tuple ): __a : List[Any] = MobileBertModelTester(self ) __a : int = ConfigTester(self , config_class=snake_case_ , hidden_size=3_7 ) def lowerCAmelCase (self : Union[str, Any] ): self.config_tester.run_common_tests() def lowerCAmelCase (self : Optional[Any] ): __a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCAmelCase (self : str ): __a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCAmelCase (self : Tuple ): __a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCAmelCase (self : Dict ): __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCAmelCase (self : int ): __a : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCAmelCase (self : List[Any] ): __a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCAmelCase (self : int ): __a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCAmelCase (self : Tuple ): __a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) def __UpperCamelCase ( lowerCAmelCase__ : str ): return torch.tensor( lowerCAmelCase__ , dtype=torch.long , device=lowerCAmelCase__ , ) lowercase__ =1e-3 @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase (self : Any ): __a : Dict = MobileBertModel.from_pretrained('''google/mobilebert-uncased''' ).to(snake_case_ ) __a : Tuple = _long_tensor([[1_0_1, 7_1_1_0, 1_0_0_5, 1_0_5_6, 2_0_2_3, 1_1_3_3_3, 1_7_4_1_3, 1_0_2_9, 1_0_2]] ) with torch.no_grad(): __a : str = model(snake_case_ )[0] __a : List[Any] = torch.Size((1, 9, 5_1_2) ) self.assertEqual(output.shape , snake_case_ ) __a : Union[str, Any] = torch.tensor( [ [ [-2.473_6526E07, 8.269_1656E04, 1.652_1838E05], [-5.754_1704E-01, 3.905_6022E00, 4.401_1507E00], [2.604_7359E00, 1.567_7652E00, -1.732_4188E-01], ] ] , device=snake_case_ , ) # MobileBERT results range from 10e0 to 10e8. Even a 0.0000001% difference with a value of 10e8 results in a # ~1 difference, it's therefore not a good idea to measure using addition. # Here, we instead divide the expected result with the result in order to obtain ~1. We then check that the # result is held between bounds: 1 - TOLERANCE < expected_result / result < 1 + TOLERANCE __a : List[str] = torch.all((expected_slice / output[..., :3, :3]) >= 1 - TOLERANCE ) __a : Any = torch.all((expected_slice / output[..., :3, :3]) <= 1 + TOLERANCE ) self.assertTrue(lower_bound and upper_bound )
216
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import copy import importlib.metadata import json import os from dataclasses import dataclass from typing import Any, Dict, Union from packaging import version from ..utils import is_torch_available, logging if is_torch_available(): import torch lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) @dataclass class __lowercase : """simple docstring""" def __init__( self : Dict , lowerCAmelCase__ : Tuple=False , lowerCAmelCase__ : Union[str, Any]=False , lowerCAmelCase__ : Any=6.0 , lowerCAmelCase__ : Optional[int]=None , lowerCAmelCase__ : List[Any]=False , lowerCAmelCase__ : Tuple=False , lowerCAmelCase__ : Optional[Any]=None , lowerCAmelCase__ : Optional[Any]="fp4" , lowerCAmelCase__ : Optional[Any]=False , **lowerCAmelCase__ : Dict , ): SCREAMING_SNAKE_CASE_: Optional[int] = load_in_abit SCREAMING_SNAKE_CASE_: Optional[Any] = load_in_abit SCREAMING_SNAKE_CASE_: Tuple = llm_inta_threshold SCREAMING_SNAKE_CASE_: str = llm_inta_skip_modules SCREAMING_SNAKE_CASE_: List[str] = llm_inta_enable_fpaa_cpu_offload SCREAMING_SNAKE_CASE_: List[Any] = llm_inta_has_fpaa_weight SCREAMING_SNAKE_CASE_: Any = bnb_abit_quant_type SCREAMING_SNAKE_CASE_: str = bnb_abit_use_double_quant if bnb_abit_compute_dtype is None: SCREAMING_SNAKE_CASE_: Optional[Any] = torch.floataa elif isinstance(lowerCAmelCase__ , lowerCAmelCase__): SCREAMING_SNAKE_CASE_: Optional[int] = getattr(lowerCAmelCase__ , lowerCAmelCase__) elif isinstance(lowerCAmelCase__ , torch.dtype): SCREAMING_SNAKE_CASE_: Optional[int] = bnb_abit_compute_dtype else: raise ValueError("bnb_4bit_compute_dtype must be a string or a torch.dtype") self.post_init() def _SCREAMING_SNAKE_CASE ( self : int): if not isinstance(self.llm_inta_threshold , lowerCAmelCase__): raise ValueError("llm_int8_threshold must be a float") if self.llm_inta_skip_modules is not None and not isinstance(self.llm_inta_skip_modules , lowerCAmelCase__): raise ValueError("llm_int8_skip_modules must be a list of strings") if not isinstance(self.llm_inta_enable_fpaa_cpu_offload , lowerCAmelCase__): raise ValueError("llm_int8_enable_fp32_cpu_offload must be a boolean") if not isinstance(self.llm_inta_has_fpaa_weight , lowerCAmelCase__): raise ValueError("llm_int8_has_fp16_weight must be a boolean") if self.bnb_abit_compute_dtype is not None and not isinstance(self.bnb_abit_compute_dtype , torch.dtype): raise ValueError("bnb_4bit_compute_dtype must be torch.dtype") if not isinstance(self.bnb_abit_quant_type , lowerCAmelCase__): raise ValueError("bnb_4bit_quant_type must be a string") if not isinstance(self.bnb_abit_use_double_quant , lowerCAmelCase__): raise ValueError("bnb_4bit_use_double_quant must be a boolean") if self.load_in_abit and not version.parse(importlib.metadata.version("bitsandbytes")) >= version.parse( "0.39.0"): raise ValueError( "4 bit quantization requires bitsandbytes>=0.39.0 - please upgrade your bitsandbytes version") def _SCREAMING_SNAKE_CASE ( self : str): return self.load_in_abit or self.load_in_abit def _SCREAMING_SNAKE_CASE ( self : str): if self.load_in_abit: return "llm_int8" elif self.load_in_abit and self.bnb_abit_quant_type == "fp4": return "fp4" elif self.load_in_abit and self.bnb_abit_quant_type == "nf4": return "nf4" else: return None @classmethod def _SCREAMING_SNAKE_CASE ( cls : Dict , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[str] , **lowerCAmelCase__ : str): SCREAMING_SNAKE_CASE_: Union[str, Any] = cls(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = [] for key, value in kwargs.items(): if hasattr(lowerCAmelCase__ , lowerCAmelCase__): setattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) to_remove.append(lowerCAmelCase__) for key in to_remove: kwargs.pop(lowerCAmelCase__ , lowerCAmelCase__) if return_unused_kwargs: return config, kwargs else: return config def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : Union[str, os.PathLike]): with open(lowerCAmelCase__ , "w" , encoding="utf-8") as writer: SCREAMING_SNAKE_CASE_: Any = self.to_dict() SCREAMING_SNAKE_CASE_: List[Any] = json.dumps(lowerCAmelCase__ , indent=2 , sort_keys=lowerCAmelCase__) + "\n" writer.write(lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Dict = copy.deepcopy(self.__dict__) SCREAMING_SNAKE_CASE_: Optional[int] = str(output["bnb_4bit_compute_dtype"]).split(".")[1] return output def __repr__( self : str): return F"{self.__class__.__name__} {self.to_json_string()}" def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase__ : bool = True): if use_diff is True: SCREAMING_SNAKE_CASE_: List[Any] = self.to_diff_dict() else: SCREAMING_SNAKE_CASE_: Dict = self.to_dict() return json.dumps(lowerCAmelCase__ , indent=2 , sort_keys=lowerCAmelCase__) + "\n" def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_: Dict = self.to_dict() # get the default config dict SCREAMING_SNAKE_CASE_: Union[str, Any] = BitsAndBytesConfig().to_dict() SCREAMING_SNAKE_CASE_: int = {} # only serialize values that differ from the default config for key, value in config_dict.items(): if value != default_config_dict[key]: SCREAMING_SNAKE_CASE_: Union[str, Any] = value return serializable_config_dict
127
import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCAmelCase : Union[str, Any] = get_tests_dir("""fixtures/dummy-config.json""") class __lowercase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Dict = 0 def _SCREAMING_SNAKE_CASE ( self : Any): self.assertIsNotNone(transformers.models.auto.__spec__) self.assertIsNotNone(importlib.util.find_spec("transformers.models.auto")) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: List[Any] = AutoConfig.from_pretrained("bert-base-uncased") self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: Optional[Any] = AutoConfig.from_pretrained(lowerCAmelCase__) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: Tuple = AutoConfig.from_pretrained(lowerCAmelCase__) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: List[Any] = AutoConfig.for_model("roberta") self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[Any]): with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. SCREAMING_SNAKE_CASE_: int = os.path.join(lowerCAmelCase__ , "fake-roberta") os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__) with open(os.path.join(lowerCAmelCase__ , "config.json") , "w") as f: f.write(json.dumps({})) SCREAMING_SNAKE_CASE_: Any = AutoConfig.from_pretrained(lowerCAmelCase__) self.assertEqual(type(lowerCAmelCase__) , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): try: AutoConfig.register("custom" , lowerCAmelCase__) # Wrong model type will raise an error with self.assertRaises(lowerCAmelCase__): AutoConfig.register("model" , lowerCAmelCase__) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCAmelCase__): AutoConfig.register("bert" , lowerCAmelCase__) # Now that the config is registered, it can be used as any other config with the auto-API SCREAMING_SNAKE_CASE_: List[Any] = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Tuple = AutoConfig.from_pretrained(lowerCAmelCase__) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def _SCREAMING_SNAKE_CASE ( self : List[str]): with self.assertRaisesRegex( lowerCAmelCase__ , "bert-base is not a local folder and is not a valid model identifier"): SCREAMING_SNAKE_CASE_: List[str] = AutoConfig.from_pretrained("bert-base") def _SCREAMING_SNAKE_CASE ( self : List[Any]): with self.assertRaisesRegex( lowerCAmelCase__ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)"): SCREAMING_SNAKE_CASE_: str = AutoConfig.from_pretrained(lowerCAmelCase__ , revision="aaaaaa") def _SCREAMING_SNAKE_CASE ( self : Optional[int]): with self.assertRaisesRegex( lowerCAmelCase__ , "hf-internal-testing/no-config-test-repo does not appear to have a file named config.json." , ): SCREAMING_SNAKE_CASE_: int = AutoConfig.from_pretrained("hf-internal-testing/no-config-test-repo") def _SCREAMING_SNAKE_CASE ( self : List[str]): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: str = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model") # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: Optional[Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__) self.assertEqual(config.__class__.__name__ , "NewModelConfig") # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = AutoConfig.from_pretrained(lowerCAmelCase__ , trust_remote_code=lowerCAmelCase__) self.assertEqual(reloaded_config.__class__.__name__ , "NewModelConfig") def _SCREAMING_SNAKE_CASE ( self : List[Any]): class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : int = '''new-model''' try: AutoConfig.register("new-model" , lowerCAmelCase__) # If remote code is not set, the default is to use local SCREAMING_SNAKE_CASE_: int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model") self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal") # If remote code is disabled, we load the local one. SCREAMING_SNAKE_CASE_: Optional[int] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal") # If remote is enabled, we load from the Hub SCREAMING_SNAKE_CASE_: int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__) self.assertEqual(config.__class__.__name__ , "NewModelConfig") finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
127
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class A ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int],lowercase_ : str,lowercase_ : Optional[Any]=7,lowercase_ : List[str]=3,lowercase_ : Optional[Any]=1_8,lowercase_ : int=3_0,lowercase_ : List[Any]=4_0_0,lowercase_ : str=True,lowercase_ : List[str]=None,lowercase_ : str=True,lowercase_ : Optional[int]=None,lowercase_ : List[Any]=True,)-> Optional[Any]: '''simple docstring''' A__ = size if size is not None else {'shortest_edge': 2_0} A__ = crop_size if crop_size is not None else {'height': 1_8, 'width': 1_8} A__ = parent A__ = batch_size A__ = num_channels A__ = image_size A__ = min_resolution A__ = max_resolution A__ = do_resize A__ = size A__ = do_center_crop A__ = crop_size A__ = do_flip_channel_order def snake_case__ ( self : Optional[Any] )-> Dict: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class A ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = MobileViTImageProcessor if is_vision_available() else None def snake_case__ ( self : Optional[Any] )-> Union[str, Any]: '''simple docstring''' A__ = MobileViTImageProcessingTester(self ) @property def snake_case__ ( self : int )-> str: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self : Optional[Any] )-> List[str]: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_,'do_resize' ) ) self.assertTrue(hasattr(lowercase_,'size' ) ) self.assertTrue(hasattr(lowercase_,'do_center_crop' ) ) self.assertTrue(hasattr(lowercase_,'center_crop' ) ) self.assertTrue(hasattr(lowercase_,'do_flip_channel_order' ) ) def snake_case__ ( self : Any )-> List[str]: '''simple docstring''' A__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size,{'shortest_edge': 2_0} ) self.assertEqual(image_processor.crop_size,{'height': 1_8, 'width': 1_8} ) A__ = self.image_processing_class.from_dict(self.image_processor_dict,size=4_2,crop_size=8_4 ) self.assertEqual(image_processor.size,{'shortest_edge': 4_2} ) self.assertEqual(image_processor.crop_size,{'height': 8_4, 'width': 8_4} ) def snake_case__ ( self : int )-> List[str]: '''simple docstring''' pass def snake_case__ ( self : Union[str, Any] )-> int: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ = prepare_image_inputs(self.image_processor_tester,equal_resolution=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,Image.Image ) # Test not batched input A__ = image_processing(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),) # Test batched A__ = image_processing(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),) def snake_case__ ( self : Optional[Any] )-> Any: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ = prepare_image_inputs(self.image_processor_tester,equal_resolution=lowercase_,numpify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,np.ndarray ) # Test not batched input A__ = image_processing(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),) # Test batched A__ = image_processing(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),) def snake_case__ ( self : List[str] )-> Tuple: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ = prepare_image_inputs(self.image_processor_tester,equal_resolution=lowercase_,torchify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,torch.Tensor ) # Test not batched input A__ = image_processing(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),) # Test batched A__ = image_processing(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ),)
7
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A : """simple docstring""" def __init__( self : Union[str, Any],lowercase_ : Any,lowercase_ : Union[str, Any]=1_3,lowercase_ : Tuple=3_0,lowercase_ : List[Any]=2,lowercase_ : Optional[int]=3,lowercase_ : Union[str, Any]=True,lowercase_ : Tuple=True,lowercase_ : Any=3_2,lowercase_ : List[str]=2,lowercase_ : Optional[int]=4,lowercase_ : Union[str, Any]=3_7,lowercase_ : Tuple="gelu",lowercase_ : str=0.1,lowercase_ : Tuple=0.1,lowercase_ : Union[str, Any]=1_0,lowercase_ : int=0.02,lowercase_ : List[Any]=3,lowercase_ : Any=None,)-> Dict: '''simple docstring''' A__ = parent A__ = batch_size A__ = image_size A__ = patch_size A__ = num_channels A__ = is_training A__ = use_labels A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = type_sequence_label_size A__ = initializer_range A__ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) A__ = (image_size // patch_size) ** 2 A__ = num_patches + 1 def snake_case__ ( self : int )-> List[str]: '''simple docstring''' A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size],self.type_sequence_label_size ) A__ = self.get_config() return config, pixel_values, labels def snake_case__ ( self : Tuple )-> List[Any]: '''simple docstring''' return ViTConfig( image_size=self.image_size,patch_size=self.patch_size,num_channels=self.num_channels,hidden_size=self.hidden_size,num_hidden_layers=self.num_hidden_layers,num_attention_heads=self.num_attention_heads,intermediate_size=self.intermediate_size,hidden_act=self.hidden_act,hidden_dropout_prob=self.hidden_dropout_prob,attention_probs_dropout_prob=self.attention_probs_dropout_prob,is_decoder=lowercase_,initializer_range=self.initializer_range,) def snake_case__ ( self : List[str],lowercase_ : int,lowercase_ : Union[str, Any],lowercase_ : Tuple )-> Optional[Any]: '''simple docstring''' A__ = TFViTModel(config=lowercase_ ) A__ = model(lowercase_,training=lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. A__ = self.image_size // 2 A__ = pixel_values[:, :, :image_size, :image_size] A__ = model(lowercase_,interpolate_pos_encoding=lowercase_,training=lowercase_ ) A__ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, seq_length, self.hidden_size) ) def snake_case__ ( self : List[Any],lowercase_ : List[Any],lowercase_ : List[Any],lowercase_ : List[Any] )-> Dict: '''simple docstring''' A__ = self.type_sequence_label_size A__ = TFViTForImageClassification(lowercase_ ) A__ = model(lowercase_,labels=lowercase_,training=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. A__ = self.image_size // 2 A__ = pixel_values[:, :, :image_size, :image_size] A__ = model(lowercase_,interpolate_pos_encoding=lowercase_,training=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images A__ = 1 A__ = TFViTForImageClassification(lowercase_ ) A__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self : Any )-> Optional[Any]: '''simple docstring''' A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def snake_case__ ( self : int )-> List[Any]: '''simple docstring''' A__ = TFViTModelTester(self ) A__ = ConfigTester(self,config_class=lowercase_,has_text_modality=lowercase_,hidden_size=3_7 ) def snake_case__ ( self : Any )-> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def snake_case__ ( self : Optional[Any] )-> str: '''simple docstring''' pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def snake_case__ ( self : Any )-> int: '''simple docstring''' pass def snake_case__ ( self : str )-> Dict: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings(),(tf.keras.layers.Layer) ) A__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_,tf.keras.layers.Layer ) ) def snake_case__ ( self : int )-> List[str]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(lowercase_ ) A__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ['pixel_values'] self.assertListEqual(arg_names[:1],lowercase_ ) def snake_case__ ( self : Union[str, Any] )-> Optional[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def snake_case__ ( self : Optional[Any] )-> Optional[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) @slow def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' A__ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(lowercase_ ) def _snake_case( ) -> str: '''simple docstring''' A__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : List[Any] )-> str: '''simple docstring''' return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def snake_case__ ( self : Any )-> Dict: '''simple docstring''' A__ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=lowercase_,return_tensors='tf' ) # forward pass A__ = model(**lowercase_ ) # verify the logits A__ = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape,lowercase_ ) A__ = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3],lowercase_,atol=1E-4 )
7
1
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : int ) -> int: if n == 1 or not isinstance(__UpperCamelCase , __UpperCamelCase ): return 0 elif n == 2: return 1 else: UpperCAmelCase_ = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def SCREAMING_SNAKE_CASE ( __UpperCamelCase : int ) -> int: UpperCAmelCase_ = 0 UpperCAmelCase_ = 2 while digits < n: index += 1 UpperCAmelCase_ = len(str(fibonacci(__UpperCamelCase ) ) ) return index def SCREAMING_SNAKE_CASE ( __UpperCamelCase : int = 1000 ) -> int: return fibonacci_digits_index(__UpperCamelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
177
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) _lowerCamelCase = logging.getLogger(__name__) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser( description='Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)' ) parser.add_argument( '--data_file', type=str, default='data/dump.bert-base-uncased.pickle', help='The binarized dataset.' ) parser.add_argument( '--token_counts_dump', type=str, default='data/token_counts.bert-base-uncased.pickle', help='The dump file.' ) parser.add_argument('--vocab_size', default=3_05_22, type=int) _lowerCamelCase = parser.parse_args() logger.info(F"Loading data from {args.data_file}") with open(args.data_file, 'rb') as fp: _lowerCamelCase = pickle.load(fp) logger.info('Counting occurrences for MLM.') _lowerCamelCase = Counter() for tk_ids in data: counter.update(tk_ids) _lowerCamelCase = [0] * args.vocab_size for k, v in counter.items(): _lowerCamelCase = v logger.info(F"Dump to {args.token_counts_dump}") with open(args.token_counts_dump, 'wb') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
177
1
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Optional[int] = logging.get_logger(__name__) # TODO Update this UpperCAmelCase : Optional[Any] = { """facebook/esm-1b""": """https://huggingface.co/facebook/esm-1b/resolve/main/config.json""", # See all ESM models at https://huggingface.co/models?filter=esm } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """esm""" def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=7_6_8 , lowerCAmelCase__=1_2 , lowerCAmelCase__=1_2 , lowerCAmelCase__=3_0_7_2 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_0_2_6 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__="absolute" , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> List[Any]: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , mask_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a__ : Union[str, Any] =vocab_size a__ : List[Any] =hidden_size a__ : Optional[Any] =num_hidden_layers a__ : str =num_attention_heads a__ : Tuple =intermediate_size a__ : List[Any] =hidden_dropout_prob a__ : Optional[int] =attention_probs_dropout_prob a__ : int =max_position_embeddings a__ : List[str] =initializer_range a__ : Optional[int] =layer_norm_eps a__ : Dict =position_embedding_type a__ : int =use_cache a__ : Tuple =emb_layer_norm_before a__ : Union[str, Any] =token_dropout a__ : List[Any] =is_folding_model if is_folding_model: if esmfold_config is None: logger.info("No esmfold_config supplied for folding model, using default values." ) a__ : List[Any] =EsmFoldConfig() elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a__ : Any =EsmFoldConfig(**lowerCAmelCase__ ) a__ : Any =esmfold_config if vocab_list is None: logger.warning("No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!" ) a__ : List[Any] =get_default_vocab_list() else: a__ : List[str] =vocab_list else: a__ : Any =None a__ : str =None if self.esmfold_config is not None and getattr(self.esmfold_config , "use_esm_attn_map" , lowerCAmelCase__ ): raise ValueError("The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!" ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : Optional[Any] =super().to_dict() if isinstance(self.esmfold_config , lowerCAmelCase__ ): a__ : List[str] =self.esmfold_config.to_dict() return output @dataclass class __lowerCAmelCase : _lowercase : str = None _lowercase : bool = True _lowercase : bool = False _lowercase : bool = False _lowercase : bool = False _lowercase : float = 0 _lowercase : bool = True _lowercase : bool = False _lowercase : int = 128 _lowercase : "TrunkConfig" = None def _lowercase ( self ) -> List[Any]: '''simple docstring''' if self.trunk is None: a__ : List[Any] =TrunkConfig() elif isinstance(self.trunk , lowerCAmelCase__ ): a__ : Dict =TrunkConfig(**self.trunk ) def _lowercase ( self ) -> Dict: '''simple docstring''' a__ : Any =asdict(self ) a__ : List[str] =self.trunk.to_dict() return output @dataclass class __lowerCAmelCase : _lowercase : int = 48 _lowercase : int = 1024 _lowercase : int = 128 _lowercase : int = 32 _lowercase : int = 32 _lowercase : int = 32 _lowercase : float = 0 _lowercase : float = 0 _lowercase : bool = False _lowercase : int = 4 _lowercase : Optional[int] = 128 _lowercase : "StructureModuleConfig" = None def _lowercase ( self ) -> Any: '''simple docstring''' if self.structure_module is None: a__ : List[Any] =StructureModuleConfig() elif isinstance(self.structure_module , lowerCAmelCase__ ): a__ : Dict =StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F'''`max_recycles` should be positive, got {self.max_recycles}.''' ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( "`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got" F''' {self.sequence_state_dim} and {self.sequence_state_dim}.''' ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( "`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got" F''' {self.pairwise_state_dim} and {self.pairwise_state_dim}.''' ) a__ : int =self.sequence_state_dim // self.sequence_head_width a__ : Dict =self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( "`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got" F''' {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.''' ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( "`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got" F''' {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.''' ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F'''`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.''' ) if self.dropout >= 0.4: raise ValueError(F'''`dropout` should not be greater than 0.4, got {self.dropout}.''' ) def _lowercase ( self ) -> Any: '''simple docstring''' a__ : str =asdict(self ) a__ : List[str] =self.structure_module.to_dict() return output @dataclass class __lowerCAmelCase : _lowercase : int = 384 _lowercase : int = 128 _lowercase : int = 16 _lowercase : int = 128 _lowercase : int = 12 _lowercase : int = 4 _lowercase : int = 8 _lowercase : float = 0.1 _lowercase : int = 8 _lowercase : int = 1 _lowercase : int = 2 _lowercase : int = 7 _lowercase : int = 10 _lowercase : float = 1E-8 _lowercase : float = 1E5 def _lowercase ( self ) -> Tuple: '''simple docstring''' return asdict(self ) def _A ( ): """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
95
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __a : @staticmethod def SCREAMING_SNAKE_CASE__ ( *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: '''simple docstring''' pass def snake_case_ ( snake_case ) -> Optional[Any]: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. __lowerCAmelCase = ( '''https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png''' ) @is_pipeline_test @require_torch @require_vision class __a ( unittest.TestCase ): __lowercase : Dict = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' lowercase__: Optional[Any] = pipeline( 'document-question-answering' , model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) lowercase__: int = INVOICE_URL lowercase__: Dict = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) lowercase__: str = 'What is the placebo?' lowercase__: Any = [ { 'image': load_image(lowerCAmelCase__ ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' lowercase__: str = dqa_pipeline(lowerCAmelCase__ , top_k=2 ) self.assertEqual( lowerCAmelCase__ , [ [ {'score': ANY(lowerCAmelCase__ ), 'answer': ANY(lowerCAmelCase__ ), 'start': ANY(lowerCAmelCase__ ), 'end': ANY(lowerCAmelCase__ )}, {'score': ANY(lowerCAmelCase__ ), 'answer': ANY(lowerCAmelCase__ ), 'start': ANY(lowerCAmelCase__ ), 'end': ANY(lowerCAmelCase__ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: Union[str, Any] = pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) lowercase__: Optional[Any] = INVOICE_URL lowercase__: int = 'How many cats are there?' lowercase__: List[str] = [ {'score': 0.0_0_0_1, 'answer': 'oy 2312/2019', 'start': 38, 'end': 39}, {'score': 0.0_0_0_1, 'answer': 'oy 2312/2019 DUE', 'start': 38, 'end': 40}, ] lowercase__: Dict = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) lowercase__: Tuple = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowercase__: str = './tests/fixtures/tests_samples/COCO/000000039769.png' lowercase__: Tuple = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) # We can optionnally pass directly the words and bounding boxes lowercase__: int = './tests/fixtures/tests_samples/COCO/000000039769.png' lowercase__: List[Any] = [] lowercase__: Optional[int] = [] lowercase__: Any = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , words=lowerCAmelCase__ , boxes=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' lowercase__: List[str] = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) lowercase__: int = INVOICE_URL lowercase__: str = 'What is the invoice number?' lowercase__: Union[str, Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Any = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Optional[int] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: Any = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=50 , ) lowercase__: Optional[int] = INVOICE_URL lowercase__: Union[str, Any] = 'What is the invoice number?' lowercase__: Optional[Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Tuple = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Dict = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: Optional[Any] = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCAmelCase__ ) lowercase__: Optional[Any] = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCAmelCase__ , revision='3dc6de3' , ) lowercase__: List[str] = INVOICE_URL lowercase__: Union[str, Any] = 'What is the invoice number?' lowercase__: Dict = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowercase__: List[str] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowercase__: int = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] ] * 2 , ) lowercase__: Any = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) # This model should also work if `image` is set to None lowercase__: List[Any] = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: Any = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCAmelCase__ ) lowercase__: str = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCAmelCase__ , revision='3dc6de3' , max_seq_len=50 , ) lowercase__: Optional[Any] = INVOICE_URL lowercase__: Optional[Any] = 'What is the invoice number?' lowercase__: Optional[int] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Any = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) lowercase__: Optional[int] = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) # This model should also work if `image` is set to None lowercase__: Tuple = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) @slow @require_torch def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' lowercase__: List[Any] = pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) lowercase__: int = INVOICE_URL lowercase__: int = 'What is the invoice number?' lowercase__: Union[str, Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' pass
196
0
"""simple docstring""" from __future__ import annotations def __lowercase ( snake_case_ : list[list[int]] ) ->int: '''simple docstring''' for i in range(1 ,len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 ,len(snake_case_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 ,len(snake_case_ ) ): for j in range(1 ,len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] ,matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
291
"""simple docstring""" import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = MgpstrTokenizer _lowerCamelCase = False _lowerCamelCase = {} _lowerCamelCase = False def UpperCamelCase__( self ): '''simple docstring''' super().setUp() # fmt: off __A : int = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on __A : Dict = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) __A : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowerCamelCase ) + '''\n''' ) def UpperCamelCase__( self , **__lowerCamelCase ): '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : List[str] = '''tester''' __A : Dict = '''tester''' return input_text, output_text @unittest.skip('''MGP-STR always lower cases letters.''' ) def UpperCamelCase__( self ): '''simple docstring''' pass def UpperCamelCase__( self ): '''simple docstring''' __A : List[Any] = self.get_tokenizers(do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): __A : Union[str, Any] = '''[SPECIAL_TOKEN]''' tokenizer.add_special_tokens({'''cls_token''': special_token} ) __A : Optional[Any] = tokenizer.encode([special_token] , add_special_tokens=__lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , 1 ) __A : List[Any] = tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) self.assertTrue(special_token not in decoded ) def UpperCamelCase__( self ): '''simple docstring''' __A : Tuple = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): __A , __A : str = self.get_input_output_texts(__lowerCamelCase ) __A : Union[str, Any] = tokenizer.tokenize(__lowerCamelCase ) __A : Union[str, Any] = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) __A : Union[str, Any] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) __A : Optional[Any] = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertNotEqual(len(__lowerCamelCase ) , 0 ) __A : Union[str, Any] = tokenizer.decode(__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertEqual(text_a.replace(''' ''' , '''''' ) , __lowerCamelCase ) @unittest.skip('''MGP-STR tokenizer only handles one sequence.''' ) def UpperCamelCase__( self ): '''simple docstring''' pass @unittest.skip('''inputs cannot be pretokenized in MgpstrTokenizer''' ) def UpperCamelCase__( self ): '''simple docstring''' pass
291
1
"""simple docstring""" lowerCamelCase__ = "Input must be a string of 8 numbers plus letter" lowerCamelCase__ = "TRWAGMYFPDXBNJZSQVHLCKE" def __lowerCAmelCase (_UpperCamelCase ): if not isinstance(_snake_case , _snake_case ): __lowerCAmelCase : List[str] = F"Expected string as input, found {type(_snake_case ).__name__}" raise TypeError(_snake_case ) __lowerCAmelCase : int = spanish_id.replace('-' , '' ).upper() if len(_snake_case ) != 9: raise ValueError(_snake_case ) try: __lowerCAmelCase : Optional[int] = int(spanish_id_clean[0:8] ) __lowerCAmelCase : Optional[int] = spanish_id_clean[8] except ValueError as ex: raise ValueError(_snake_case ) from ex if letter.isdigit(): raise ValueError(_snake_case ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
86
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if isinstance(_snake_case , _snake_case ): __magic_name__ : Union[str, Any] = np.full((len(_snake_case ), sequence_length, 2) , _snake_case ) else: __magic_name__ : List[Any] = np.full((len(_snake_case ), sequence_length) , _snake_case ) for i, tensor in enumerate(_snake_case ): if padding_side == "right": if isinstance(_snake_case , _snake_case ): __magic_name__ : Optional[Any] = tensor[:sequence_length] else: __magic_name__ : Union[str, Any] = tensor[:sequence_length] else: if isinstance(_snake_case , _snake_case ): __magic_name__ : List[Any] = tensor[:sequence_length] else: __magic_name__ : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def lowerCAmelCase_ ( _snake_case : Optional[int] ) -> Tuple: '''simple docstring''' __magic_name__ : Union[str, Any] = ord(_snake_case ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True __magic_name__ : Any = unicodedata.category(_snake_case ) if cat.startswith("P" ): return True return False @dataclass class _snake_case ( snake_case ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -100 UpperCamelCase__ = "pt" def SCREAMING_SNAKE_CASE ( self , _a ): import torch __magic_name__ : List[str] = "label" if "label" in features[0].keys() else "labels" __magic_name__ : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __magic_name__ : Optional[int] = self.tokenizer.pad( _a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" if labels is None else None , ) if labels is None: return batch __magic_name__ : Dict = torch.tensor(batch["entity_ids"] ).shape[1] __magic_name__ : List[Any] = self.tokenizer.padding_side if padding_side == "right": __magic_name__ : str = [ list(_a ) + [self.label_pad_token_id] * (sequence_length - len(_a )) for label in labels ] else: __magic_name__ : int = [ [self.label_pad_token_id] * (sequence_length - len(_a )) + list(_a ) for label in labels ] __magic_name__ : Dict = [feature["ner_tags"] for feature in features] __magic_name__ : List[Any] = padding_tensor(_a , -1 , _a , _a ) __magic_name__ : Any = [feature["original_entity_spans"] for feature in features] __magic_name__ : Any = padding_tensor(_a , (-1, -1) , _a , _a ) __magic_name__ : List[Any] = {k: torch.tensor(_a , dtype=torch.intaa ) for k, v in batch.items()} return batch
281
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: return 1 if input_a == input_a else 0 def lowerCAmelCase__ ( ) -> None: assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
307
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = { 'task_specific_params': { 'summarization': {'length_penalty': 1.0, 'max_length': 128, 'min_length': 12, 'num_beams': 4}, 'summarization_cnn': {'length_penalty': 2.0, 'max_length': 142, 'min_length': 56, 'num_beams': 4}, 'summarization_xsum': {'length_penalty': 1.0, 'max_length': 62, 'min_length': 11, 'num_beams': 6}, } } lowerCAmelCase__ : int = { 'task_specific_params.summarization.length_penalty': 1.0, 'task_specific_params.summarization.max_length': 128, 'task_specific_params.summarization.min_length': 12, 'task_specific_params.summarization.num_beams': 4, 'task_specific_params.summarization_cnn.length_penalty': 2.0, 'task_specific_params.summarization_cnn.max_length': 142, 'task_specific_params.summarization_cnn.min_length': 56, 'task_specific_params.summarization_cnn.num_beams': 4, 'task_specific_params.summarization_xsum.length_penalty': 1.0, 'task_specific_params.summarization_xsum.max_length': 62, 'task_specific_params.summarization_xsum.min_length': 11, 'task_specific_params.summarization_xsum.num_beams': 6, } self.assertEqual(flatten_dict(a ) , a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(a ) , x.transpose() ) ) lowerCAmelCase__ : List[str] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = np.random.randn(3 , 4 ) lowerCAmelCase__ : List[Any] = torch.tensor(a ) self.assertTrue(np.allclose(transpose(a ) , transpose(a ).numpy() ) ) lowerCAmelCase__ : str = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : int = torch.tensor(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , transpose(a , axes=(1, 2, 0) ).numpy() ) ) @require_tf def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = np.random.randn(3 , 4 ) lowerCAmelCase__ : Any = tf.constant(a ) self.assertTrue(np.allclose(transpose(a ) , transpose(a ).numpy() ) ) lowerCAmelCase__ : str = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : Dict = tf.constant(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , transpose(a , axes=(1, 2, 0) ).numpy() ) ) @require_flax def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = np.random.randn(3 , 4 ) lowerCAmelCase__ : int = jnp.array(a ) self.assertTrue(np.allclose(transpose(a ) , np.asarray(transpose(a ) ) ) ) lowerCAmelCase__ : Any = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : str = jnp.array(a ) self.assertTrue(np.allclose(transpose(a , axes=(1, 2, 0) ) , np.asarray(transpose(a , axes=(1, 2, 0) ) ) ) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , np.reshape(a , (4, 3) ) ) ) lowerCAmelCase__ : Tuple = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , np.reshape(a , (12, 5) ) ) ) @require_torch def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = np.random.randn(3 , 4 ) lowerCAmelCase__ : Dict = torch.tensor(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , reshape(a , (4, 3) ).numpy() ) ) lowerCAmelCase__ : str = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : str = torch.tensor(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , reshape(a , (12, 5) ).numpy() ) ) @require_tf def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = np.random.randn(3 , 4 ) lowerCAmelCase__ : List[Any] = tf.constant(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , reshape(a , (4, 3) ).numpy() ) ) lowerCAmelCase__ : Dict = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : Any = tf.constant(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , reshape(a , (12, 5) ).numpy() ) ) @require_flax def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = np.random.randn(3 , 4 ) lowerCAmelCase__ : List[str] = jnp.array(a ) self.assertTrue(np.allclose(reshape(a , (4, 3) ) , np.asarray(reshape(a , (4, 3) ) ) ) ) lowerCAmelCase__ : str = np.random.randn(3 , 4 , 5 ) lowerCAmelCase__ : Union[str, Any] = jnp.array(a ) self.assertTrue(np.allclose(reshape(a , (12, 5) ) , np.asarray(reshape(a , (12, 5) ) ) ) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(a ) , np.squeeze(a ) ) ) lowerCAmelCase__ : int = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , np.squeeze(a , axis=2 ) ) ) @require_torch def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = np.random.randn(1 , 3 , 4 ) lowerCAmelCase__ : str = torch.tensor(a ) self.assertTrue(np.allclose(squeeze(a ) , squeeze(a ).numpy() ) ) lowerCAmelCase__ : Optional[Any] = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase__ : Dict = torch.tensor(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , squeeze(a , axis=2 ).numpy() ) ) @require_tf def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = np.random.randn(1 , 3 , 4 ) lowerCAmelCase__ : Any = tf.constant(a ) self.assertTrue(np.allclose(squeeze(a ) , squeeze(a ).numpy() ) ) lowerCAmelCase__ : int = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase__ : str = tf.constant(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , squeeze(a , axis=2 ).numpy() ) ) @require_flax def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = np.random.randn(1 , 3 , 4 ) lowerCAmelCase__ : Union[str, Any] = jnp.array(a ) self.assertTrue(np.allclose(squeeze(a ) , np.asarray(squeeze(a ) ) ) ) lowerCAmelCase__ : str = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase__ : Optional[Any] = jnp.array(a ) self.assertTrue(np.allclose(squeeze(a , axis=2 ) , np.asarray(squeeze(a , axis=2 ) ) ) ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , np.expand_dims(a , axis=1 ) ) ) @require_torch def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : str = np.random.randn(3 , 4 ) lowerCAmelCase__ : str = torch.tensor(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , expand_dims(a , axis=1 ).numpy() ) ) @require_tf def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = np.random.randn(3 , 4 ) lowerCAmelCase__ : Any = tf.constant(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , expand_dims(a , axis=1 ).numpy() ) ) @require_flax def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = np.random.randn(3 , 4 ) lowerCAmelCase__ : Tuple = jnp.array(a ) self.assertTrue(np.allclose(expand_dims(a , axis=1 ) , np.asarray(expand_dims(a , axis=1 ) ) ) )
307
1
"""simple docstring""" import argparse import json from tqdm import tqdm def a_ ( ): UpperCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--src_path' , type=lowerCamelCase , default='biencoder-nq-dev.json' , help='Path to raw DPR training data' , ) parser.add_argument( '--evaluation_set' , type=lowerCamelCase , help='where to store parsed evaluation_set file' , ) parser.add_argument( '--gold_data_path' , type=lowerCamelCase , help='where to store parsed gold_data_path file' , ) UpperCAmelCase__ = parser.parse_args() with open(args.src_path , 'r' ) as src_file, open(args.evaluation_set , 'w' ) as eval_file, open( args.gold_data_path , 'w' ) as gold_file: UpperCAmelCase__ = json.load(lowerCamelCase ) for dpr_record in tqdm(lowerCamelCase ): UpperCAmelCase__ = dpr_record['question'] UpperCAmelCase__ = [context['title'] for context in dpr_record['positive_ctxs']] eval_file.write(question + '\n' ) gold_file.write('\t'.join(lowerCamelCase ) + '\n' ) if __name__ == "__main__": main()
98
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase : Optional[int] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[str] = { '''configuration_trajectory_transformer''': [ '''TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrajectoryTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[Any] = [ '''TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrajectoryTransformerModel''', '''TrajectoryTransformerPreTrainedModel''', '''load_tf_weights_in_trajectory_transformer''', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys a__ : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
195
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a__ : List[str] = { '''configuration_biogpt''': ['''BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BioGptConfig'''], '''tokenization_biogpt''': ['''BioGptTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Union[str, Any] = [ '''BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BioGptForCausalLM''', '''BioGptForTokenClassification''', '''BioGptForSequenceClassification''', '''BioGptModel''', '''BioGptPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys a__ : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
195
1
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=3 , A_=32 , A_=3 , A_=10 , A_=[10, 20, 30, 40] , A_=[1, 1, 2, 1] , A_=True , A_=True , A_="relu" , A_=3 , A_=None , ): '''simple docstring''' UpperCamelCase : List[Any] = parent UpperCamelCase : Optional[Any] = batch_size UpperCamelCase : Union[str, Any] = image_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Any = embeddings_size UpperCamelCase : Any = hidden_sizes UpperCamelCase : List[str] = depths UpperCamelCase : Optional[int] = is_training UpperCamelCase : Optional[Any] = use_labels UpperCamelCase : Dict = hidden_act UpperCamelCase : List[str] = num_labels UpperCamelCase : str = scope UpperCamelCase : Optional[Any] = len(lowerCamelCase__ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : Any = self.get_config() return config, pixel_values def __UpperCamelCase( self ): '''simple docstring''' return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = FlaxRegNetModel(config=lowerCamelCase__ ) UpperCamelCase : str = model(lowerCamelCase__ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.num_labels UpperCamelCase : Union[str, Any] = FlaxRegNetForImageClassification(config=lowerCamelCase__ ) UpperCamelCase : Union[str, Any] = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.prepare_config_and_inputs() UpperCamelCase : List[Any] = config_and_inputs UpperCamelCase : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class A__ ( __lowerCAmelCase , unittest.TestCase ): _UpperCAmelCase :int = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () _UpperCAmelCase :Dict = False _UpperCAmelCase :int = False _UpperCAmelCase :Tuple = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = FlaxRegNetModelTester(self ) UpperCamelCase : str = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def __UpperCamelCase( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __UpperCamelCase( self ): '''simple docstring''' return def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @unittest.skip(reason="RegNet does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass @unittest.skip(reason="RegNet does not support input and output embeddings" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[int] = model_class(lowerCamelCase__ ) UpperCamelCase : Optional[Any] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : Tuple = [*signature.parameters.keys()] UpperCamelCase : Optional[int] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def __UpperCamelCase( self ): '''simple docstring''' def check_hidden_states_output(A_ , A_ , A_ ): UpperCamelCase : Any = model_class(lowerCamelCase__ ) UpperCamelCase : List[str] = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) UpperCamelCase : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase : Optional[Any] = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Tuple = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase : List[Any] = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCamelCase : List[str] = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase : str = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(A_ , **A_ ): return model(pixel_values=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest("JIT Enabled" ): UpperCamelCase : List[Any] = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): UpperCamelCase : List[str] = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def A_ ( ) -> List[Any]: UpperCamelCase : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_flax class A__ ( unittest.TestCase ): @cached_property def __UpperCamelCase( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained("facebook/regnet-y-040" ) if is_vision_available() else None @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = FlaxRegNetForImageClassification.from_pretrained("facebook/regnet-y-040" ) UpperCamelCase : int = self.default_image_processor UpperCamelCase : Union[str, Any] = prepare_img() UpperCamelCase : Any = image_processor(images=lowerCamelCase__ , return_tensors="np" ) UpperCamelCase : List[str] = model(**lowerCamelCase__ ) # verify the logits UpperCamelCase : Optional[Any] = (1, 1000) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) UpperCamelCase : Dict = jnp.array([-0.41_80, -1.50_51, -3.48_36] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
52
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __magic_name__ ( datasets.BeamBasedBuilder): def UpperCAmelCase__ ( self : Tuple ) -> List[Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({'''content''': datasets.Value('''string''' )} ) , supervised_keys=lowerCamelCase__ , ) def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : str , lowerCamelCase__ : List[str] , lowerCamelCase__ : Dict ) -> str: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase__ ) class __magic_name__ ( datasets.BeamBasedBuilder): def UpperCAmelCase__ ( self : List[str] ) -> Any: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) , supervised_keys=lowerCamelCase__ , ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : Dict ) -> int: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : str , lowerCamelCase__ : Union[str, Any] ) -> List[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase__ ) def _a ( ): """simple docstring""" return [(i, {"content": content}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] def _a ( ): """simple docstring""" return [(i, {"a": {"b": [content]}}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] class __magic_name__ ( __lowerCAmelCase): @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[str] = DummyBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCamelCase__ : Any = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) self.assertDictEqual(dset['''train'''][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' import apache_beam as beam UpperCamelCase__ : List[Any] = beam.io.parquetio.WriteToParquet UpperCamelCase__ : str = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[Any] = DummyBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) with patch('''apache_beam.io.parquetio.WriteToParquet''' ) as write_parquet_mock: UpperCamelCase__ : Any = partial(lowerCamelCase__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCamelCase__ : Tuple = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset['''train''']['''content'''] ) , sorted(['''foo''', '''bar''', '''foobar'''] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Any ) -> Dict: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : List[Any] = DummyBeamDataset(cache_dir=lowerCamelCase__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCamelCase__ : Tuple = NestedBeamDataset(cache_dir=lowerCamelCase__ , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) ) UpperCamelCase__ : List[Any] = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , lowerCamelCase__ ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , lowerCamelCase__ ) self.assertDictEqual(dset['''train'''][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase__ , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset
146
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase__ : List[Any] = logging.get_logger(__name__) UpperCamelCase__ : Any = { '''microsoft/resnet-50''': '''https://huggingface.co/microsoft/resnet-50/blob/main/config.json''', } class _lowerCAmelCase ( lowerCamelCase__, lowerCamelCase__ ): """simple docstring""" lowerCamelCase = '''resnet''' lowerCamelCase = ['''basic''', '''bottleneck'''] def __init__( self , _lowerCamelCase=3 , _lowerCamelCase=64 , _lowerCamelCase=[256, 512, 1024, 2048] , _lowerCamelCase=[3, 4, 6, 3] , _lowerCamelCase="bottleneck" , _lowerCamelCase="relu" , _lowerCamelCase=False , _lowerCamelCase=None , _lowerCamelCase=None , **_lowerCamelCase , ) -> str: super().__init__(**_lowerCamelCase ) if layer_type not in self.layer_types: raise ValueError(F"layer_type={layer_type} is not one of {','.join(self.layer_types )}" ) A_ : List[Any] = num_channels A_ : str = embedding_size A_ : Dict = hidden_sizes A_ : str = depths A_ : Tuple = layer_type A_ : Dict = hidden_act A_ : Optional[int] = downsample_in_first_stage A_ : Optional[int] = ["stem"] + [F"stage{idx}" for idx in range(1 , len(_lowerCamelCase ) + 1 )] A_ : Optional[Any] = get_aligned_output_features_output_indices( out_features=_lowerCamelCase , out_indices=_lowerCamelCase , stage_names=self.stage_names ) class _lowerCAmelCase ( lowerCamelCase__ ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> List[str]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> List[str]: return 1e-3
354
'''simple docstring''' import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def UpperCAmelCase ( a_ ) -> Any: """simple docstring""" if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase ) -> str: super().__init__() A_ : Any = module A_ : Any = nn.Sequential( nn.Linear(module.in_features , _lowerCamelCase , bias=_lowerCamelCase ) , nn.Linear(_lowerCamelCase , module.out_features , bias=_lowerCamelCase ) , ) A_ : Union[str, Any] = (2.0 / (5 * min(module.in_features , module.out_features ))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=_lowerCamelCase ) nn.init.zeros_(self.adapter[1].weight ) self.adapter.to(module.weight.device ) def UpperCAmelCase_ ( self , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> List[Any]: return self.module(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) + self.adapter(_lowerCamelCase ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" lowerCamelCase = '''bigscience/bloom-1b7''' # Constant values lowerCamelCase = 2.1_0_9_6_5_9_5_5_2_6_9_2_5_7_4 lowerCamelCase = '''Hello my name is''' lowerCamelCase = set() EXPECTED_OUTPUTS.add('''Hello my name is John and I am a professional photographer. I''' ) EXPECTED_OUTPUTS.add('''Hello my name is John.\nI am a friend of your father.\n''' ) EXPECTED_OUTPUTS.add('''Hello my name is John Doe, I am a student at the University''' ) lowerCamelCase = 10 def UpperCAmelCase_ ( self ) -> List[str]: # Models and tokenizer A_ : List[str] = AutoTokenizer.from_pretrained(self.model_name ) class _lowerCAmelCase ( __A ): """simple docstring""" def UpperCAmelCase_ ( self ) -> Optional[Any]: super().setUp() # Models and tokenizer A_ : List[str] = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : List[Any] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ) -> List[Any]: A_ : str = self.model_abit.config self.assertTrue(hasattr(_lowerCamelCase , """quantization_config""" ) ) A_ : Union[str, Any] = config.to_dict() A_ : Optional[int] = config.to_diff_dict() A_ : Tuple = config.to_json_string() def UpperCAmelCase_ ( self ) -> str: from bitsandbytes.nn import Paramsabit A_ : List[Any] = self.model_fpaa.get_memory_footprint() A_ : Tuple = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE ) A_ : Union[str, Any] = get_some_linear_layer(self.model_abit ) self.assertTrue(linear.weight.__class__ == Paramsabit ) def UpperCAmelCase_ ( self ) -> List[str]: from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(_lowerCamelCase , torch.nn.Linear ): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : List[str] = self.tokenizer(self.input_text , return_tensors="""pt""" ) A_ : int = self.model_abit.generate(input_ids=encoded_input["""input_ids"""].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=_lowerCamelCase ) , self.EXPECTED_OUTPUTS ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = BitsAndBytesConfig() A_ : Tuple = True A_ : Any = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=_lowerCamelCase , device_map="""auto""" ) A_ : Optional[int] = self.tokenizer(self.input_text , return_tensors="""pt""" ) A_ : Optional[Any] = model_abit_from_config.generate( input_ids=encoded_input["""input_ids"""].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=_lowerCamelCase ) , self.EXPECTED_OUTPUTS ) def UpperCAmelCase_ ( self ) -> List[Any]: with self.assertRaises(_lowerCamelCase ), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Union[str, Any] = BitsAndBytesConfig() with self.assertRaises(_lowerCamelCase ): A_ : List[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=_lowerCamelCase , load_in_abit=_lowerCamelCase , device_map="""auto""" , bnb_abit_quant_type="""nf4""" , ) def UpperCAmelCase_ ( self ) -> str: with self.assertRaises(_lowerCamelCase ): # Tries with `str` self.model_abit.to("""cpu""" ) with self.assertRaises(_lowerCamelCase ): # Tries with a `dtype`` self.model_abit.to(torch.floataa ) with self.assertRaises(_lowerCamelCase ): # Tries with a `device` self.model_abit.to(torch.device("""cuda:0""" ) ) with self.assertRaises(_lowerCamelCase ): # Tries with a `device` self.model_abit.float() with self.assertRaises(_lowerCamelCase ): # Tries with a `device` self.model_abit.half() # Test if we did not break anything A_ : Optional[int] = self.tokenizer(self.input_text , return_tensors="""pt""" ) A_ : Tuple = self.model_fpaa.to(torch.floataa ) A_ : int = self.model_fpaa.generate(input_ids=encoded_input["""input_ids"""].to(0 ) , max_new_tokens=10 ) # Check this does not throw an error A_ : Any = self.model_fpaa.to("""cpu""" ) # Check this does not throw an error A_ : str = self.model_fpaa.half() # Check this does not throw an error A_ : Any = self.model_fpaa.float() def UpperCAmelCase_ ( self ) -> Dict: A_ : Optional[Any] = AutoModelForSeqaSeqLM.from_pretrained("""t5-small""" , load_in_abit=_lowerCamelCase , device_map="""auto""" ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @classmethod def UpperCAmelCase_ ( cls ) -> Optional[int]: A_ : Optional[int] = """t5-small""" A_ : List[str] = """google/flan-t5-small""" # flan-t5 uses dense-act instead of dense-relu-dense A_ : List[str] = AutoTokenizer.from_pretrained(cls.model_name ) A_ : Optional[Any] = """Translate in German: Hello, my dog is cute""" def UpperCAmelCase_ ( self ) -> Optional[Any]: gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ) -> Union[str, Any]: from transformers import TaForConditionalGeneration A_ : Optional[int] = TaForConditionalGeneration._keep_in_fpaa_modules A_ : Any = None # test with `t5-small` A_ : int = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) A_ : Optional[Any] = self.tokenizer(self.input_text , return_tensors="""pt""" ).to(0 ) A_ : Optional[int] = model.generate(**_lowerCamelCase ) # test with `flan-t5-small` A_ : Tuple = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) A_ : Dict = self.tokenizer(self.input_text , return_tensors="""pt""" ).to(0 ) A_ : str = model.generate(**_lowerCamelCase ) A_ : Optional[int] = modules def UpperCAmelCase_ ( self ) -> List[Any]: import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` A_ : str = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit ) ) A_ : Any = self.tokenizer(self.input_text , return_tensors="""pt""" ).to(0 ) A_ : List[Any] = model.generate(**_lowerCamelCase ) # test with `flan-t5-small` A_ : Union[str, Any] = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) A_ : int = self.tokenizer(self.input_text , return_tensors="""pt""" ).to(0 ) A_ : Optional[int] = model.generate(**_lowerCamelCase ) class _lowerCAmelCase ( __A ): """simple docstring""" def UpperCAmelCase_ ( self ) -> int: super().setUp() # model_name A_ : Dict = """bigscience/bloom-560m""" A_ : Union[str, Any] = """t5-small""" # Different types of model A_ : Optional[Any] = AutoModel.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) # Sequence classification model A_ : Dict = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) # CausalLM model A_ : Any = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) # Seq2seq model A_ : Tuple = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=_lowerCamelCase , device_map="""auto""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ) -> List[Any]: from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit ) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter ) class _lowerCAmelCase ( __A ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: super().setUp() def UpperCAmelCase_ ( self ) -> Any: del self.pipe gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : List[str] = pipeline( """text-generation""" , model=self.model_name , model_kwargs={"""device_map""": """auto""", """load_in_4bit""": True, """torch_dtype""": torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass A_ : int = self.pipe(self.input_text ) self.assertIn(pipeline_output[0]["""generated_text"""] , self.EXPECTED_OUTPUTS ) @require_torch_multi_gpu class _lowerCAmelCase ( __A ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: super().setUp() def UpperCAmelCase_ ( self ) -> str: A_ : List[str] = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=_lowerCamelCase , device_map="""balanced""" ) # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values() ) , {0, 1} ) # Check that inference pass works on the model A_ : str = self.tokenizer(self.input_text , return_tensors="""pt""" ) # Second real batch A_ : int = model_parallel.generate(input_ids=encoded_input["""input_ids"""].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=_lowerCamelCase ) , self.EXPECTED_OUTPUTS ) class _lowerCAmelCase ( __A ): """simple docstring""" def UpperCAmelCase_ ( self ) -> Tuple: A_ : Union[str, Any] = """facebook/opt-350m""" super().setUp() def UpperCAmelCase_ ( self ) -> Optional[Any]: if version.parse(importlib.metadata.version("""bitsandbytes""" ) ) < version.parse("""0.37.0""" ): return # Step 1: freeze all parameters A_ : Any = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=_lowerCamelCase ) self.assertEqual(set(model.hf_device_map.values() ) , {torch.cuda.current_device()} ) for param in model.parameters(): A_ : Optional[Any] = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability A_ : Any = param.data.to(torch.floataa ) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(_lowerCamelCase ) ): A_ : int = LoRALayer(module.q_proj , rank=16 ) A_ : Optional[int] = LoRALayer(module.k_proj , rank=16 ) A_ : Union[str, Any] = LoRALayer(module.v_proj , rank=16 ) # Step 3: dummy batch A_ : Dict = self.tokenizer("""Test batch """ , return_tensors="""pt""" ).to(0 ) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): A_ : Dict = model.forward(**_lowerCamelCase ) out.logits.norm().backward() for module in model.modules(): if isinstance(_lowerCamelCase , _lowerCamelCase ): self.assertTrue(module.adapter[1].weight.grad is not None ) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0 ) elif isinstance(_lowerCamelCase , nn.Embedding ): self.assertTrue(module.weight.grad is None ) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''gpt2-xl''' lowerCamelCase = 3.3_1_9_1_8_5_4_8_5_4_1_5_2_1_8_7
164
0
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : int) -> Tuple: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowerCamelCase ( self : Any) -> Any: """simple docstring""" _UpperCAmelCase = 1 _UpperCAmelCase = 3 _UpperCAmelCase = (32, 32) _UpperCAmelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(A) return image @property def _lowerCamelCase ( self : Union[str, Any]) -> List[str]: """simple docstring""" torch.manual_seed(0) _UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) return model @property def _lowerCamelCase ( self : Dict) -> Any: """simple docstring""" torch.manual_seed(0) _UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) return model @property def _lowerCamelCase ( self : str) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(A) @property def _lowerCamelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" def extract(*A : Tuple , **A : Tuple): class __lowerCAmelCase : def __init__( self : str) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = torch.ones([0]) def _lowerCamelCase ( self : List[Any] , A : List[str]) -> str: """simple docstring""" self.pixel_values.to(A) return self return Out() return extract def _lowerCamelCase ( self : Dict) -> List[Any]: """simple docstring""" _UpperCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.dummy_cond_unet _UpperCAmelCase = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' , clip_sample=A , set_alpha_to_one=A , ) _UpperCAmelCase = self.dummy_vae _UpperCAmelCase = self.dummy_text_encoder _UpperCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') # make sure here that pndm scheduler skips prk _UpperCAmelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = 'A painting of a squirrel eating a burger' _UpperCAmelCase = torch.Generator(device=A).manual_seed(0) _UpperCAmelCase = sd_pipe([prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type='np') _UpperCAmelCase = output.images _UpperCAmelCase = torch.Generator(device=A).manual_seed(0) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=A , )[0] _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase = np.array([0.5_7_5_6, 0.6_1_1_8, 0.5_0_0_5, 0.5_0_4_1, 0.5_4_7_1, 0.4_7_2_6, 0.4_9_7_6, 0.4_8_6_5, 0.4_8_6_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : List[Any]) -> List[str]: """simple docstring""" _UpperCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.dummy_cond_unet _UpperCAmelCase = PNDMScheduler(skip_prk_steps=A) _UpperCAmelCase = self.dummy_vae _UpperCAmelCase = self.dummy_text_encoder _UpperCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') # make sure here that pndm scheduler skips prk _UpperCAmelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = 'A painting of a squirrel eating a burger' _UpperCAmelCase = torch.Generator(device=A).manual_seed(0) _UpperCAmelCase = sd_pipe([prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type='np') _UpperCAmelCase = output.images _UpperCAmelCase = torch.Generator(device=A).manual_seed(0) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=A , )[0] _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase = np.array([0.5_1_2_5, 0.5_7_1_6, 0.4_8_2_8, 0.5_0_6_0, 0.5_6_5_0, 0.4_7_6_8, 0.5_1_8_5, 0.4_8_9_5, 0.4_9_9_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : Any) -> Optional[int]: """simple docstring""" _UpperCAmelCase = StableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-lms-pipe' , safety_checker=A) assert isinstance(A , A) assert isinstance(pipe.scheduler , A) assert pipe.safety_checker is None _UpperCAmelCase = pipe('example prompt' , num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A) _UpperCAmelCase = StableDiffusionPipeline.from_pretrained(A) # sanity check that the pipeline still works assert pipe.safety_checker is None _UpperCAmelCase = pipe('example prompt' , num_inference_steps=2).images[0] assert image is not None @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU') def _lowerCamelCase ( self : Dict) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = self.dummy_cond_unet _UpperCAmelCase = PNDMScheduler(skip_prk_steps=A) _UpperCAmelCase = self.dummy_vae _UpperCAmelCase = self.dummy_text_encoder _UpperCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') # put models in fp16 _UpperCAmelCase = unet.half() _UpperCAmelCase = vae.half() _UpperCAmelCase = bert.half() # make sure here that pndm scheduler skips prk _UpperCAmelCase = StableDiffusionPipeline( unet=A , scheduler=A , vae=A , text_encoder=A , tokenizer=A , safety_checker=A , feature_extractor=self.dummy_extractor , ) _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = 'A painting of a squirrel eating a burger' _UpperCAmelCase = sd_pipe([prompt] , num_inference_steps=2 , output_type='np').images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any]) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCAmelCase = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=A) _UpperCAmelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config) _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = ( 'portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle' ' coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with' ' anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and' ' children from bahnhof zoo, detailed ' ) _UpperCAmelCase = 40_03_66_03_46 _UpperCAmelCase = 7 # without safety guidance (sld_guidance_scale = 0) _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = [0.2_2_7_8, 0.2_2_3_1, 0.2_2_4_9, 0.2_3_3_3, 0.2_3_0_3, 0.1_8_8_5, 0.2_2_7_3, 0.2_1_4_4, 0.2_1_7_6] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 # without safety guidance (strong configuration) _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = [0.2_3_8_3, 0.2_2_7_6, 0.2_3_6, 0.2_1_9_2, 0.2_1_8_6, 0.2_0_5_3, 0.1_9_7_1, 0.1_9_0_1, 0.1_7_1_9] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=A) _UpperCAmelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config) _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = 'padme amidala taking a bath artwork, safe for work, no nudity' _UpperCAmelCase = 27_34_97_17_55 _UpperCAmelCase = 7 _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = [0.3_5_0_2, 0.3_6_2_2, 0.3_3_9_6, 0.3_6_4_2, 0.3_4_7_8, 0.3_3_1_8, 0.3_5, 0.3_3_4_8, 0.3_2_9_7] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = [0.5_5_3_1, 0.5_2_0_6, 0.4_8_9_5, 0.5_1_5_6, 0.5_1_8_2, 0.4_7_5_1, 0.4_8_0_2, 0.4_8_0_3, 0.4_4_4_3] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : List[str]) -> int: """simple docstring""" _UpperCAmelCase = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5') _UpperCAmelCase = sd_pipe.to(A) sd_pipe.set_progress_bar_config(disable=A) _UpperCAmelCase = ( 'the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.' ' leyendecker' ) _UpperCAmelCase = 10_44_35_52_34 _UpperCAmelCase = 12 _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-7 _UpperCAmelCase = torch.manual_seed(A) _UpperCAmelCase = sd_pipe( [prompt] , generator=A , guidance_scale=A , num_inference_steps=50 , output_type='np' , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _UpperCAmelCase = output.images _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = np.array([0.5_8_1_8, 0.6_2_8_5, 0.6_8_3_5, 0.6_0_1_9, 0.6_2_5, 0.6_7_5_4, 0.6_0_9_6, 0.6_3_3_4, 0.6_5_6_1]) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2
339
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=snake_case_ ) class _a ( snake_case_ ): """simple docstring""" _lowerCamelCase : str = field(default='audio-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) _lowerCamelCase : ClassVar[Features] = Features({'audio': Audio()} ) _lowerCamelCase : ClassVar[Features] = Features({'labels': ClassLabel} ) _lowerCamelCase : str = "audio" _lowerCamelCase : str = "labels" def __A ( self : str , UpperCAmelCase : List[Any] ): if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , UpperCAmelCase ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) A_ = copy.deepcopy(self ) A_ = self.label_schema.copy() A_ = features[self.label_column] A_ = label_schema return task_template @property def __A ( self : List[str] ): return { self.audio_column: "audio", self.label_column: "labels", }
312
0
'''simple docstring''' from __future__ import annotations def A (__lowerCamelCase :list[int] ): _lowerCAmelCase = len(__lowerCamelCase ) // 2 # choose the middle 3 elements _lowerCAmelCase = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
351
'''simple docstring''' def A (__lowerCamelCase :list[int] , __lowerCamelCase :list[int] ): # Check if the input is valid if not len(__lowerCamelCase ) == len(__lowerCamelCase ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = equationa _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = equationa # Calculate the determinants of the matrices _lowerCAmelCase = aa * ba - aa * ba _lowerCAmelCase = ca * ba - ca * ba _lowerCAmelCase = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowerCAmelCase = determinant_x / determinant _lowerCAmelCase = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
229
0
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCamelCase ( _A : np.ndarray , _A : np.ndarray ) ->float: """simple docstring""" return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(_A , _A ) ) ) def __UpperCamelCase ( _A : np.ndarray , _A : np.ndarray ) ->list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: lowerCamelCase_ =( """Wrong input data's dimensions... """ f'dataset : {dataset.ndim}, value_array : {value_array.ndim}' ) raise ValueError(_A ) try: if dataset.shape[1] != value_array.shape[1]: lowerCamelCase_ =( """Wrong input data's shape... """ f'dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}' ) raise ValueError(_A ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("""Wrong shape""" ) if dataset.dtype != value_array.dtype: lowerCamelCase_ =( """Input data have different datatype... """ f'dataset : {dataset.dtype}, value_array : {value_array.dtype}' ) raise TypeError(_A ) lowerCamelCase_ =[] for value in value_array: lowerCamelCase_ =euclidean(_A , dataset[0] ) lowerCamelCase_ =dataset[0].tolist() for dataset_value in dataset[1:]: lowerCamelCase_ =euclidean(_A , _A ) if dist > temp_dist: lowerCamelCase_ =temp_dist lowerCamelCase_ =dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCamelCase ( _A : np.ndarray , _A : np.ndarray ) ->float: """simple docstring""" return np.dot(_A , _A ) / (norm(_A ) * norm(_A )) if __name__ == "__main__": import doctest doctest.testmod()
154
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=512 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE="last" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , )-> Optional[Any]: lowerCamelCase_ =parent lowerCamelCase_ =batch_size lowerCamelCase_ =seq_length lowerCamelCase_ =is_training lowerCamelCase_ =use_input_lengths lowerCamelCase_ =use_token_type_ids lowerCamelCase_ =use_labels lowerCamelCase_ =gelu_activation lowerCamelCase_ =sinusoidal_embeddings lowerCamelCase_ =causal lowerCamelCase_ =asm lowerCamelCase_ =n_langs lowerCamelCase_ =vocab_size lowerCamelCase_ =n_special lowerCamelCase_ =hidden_size lowerCamelCase_ =num_hidden_layers lowerCamelCase_ =num_attention_heads lowerCamelCase_ =hidden_dropout_prob lowerCamelCase_ =attention_probs_dropout_prob lowerCamelCase_ =max_position_embeddings lowerCamelCase_ =type_vocab_size lowerCamelCase_ =type_sequence_label_size lowerCamelCase_ =initializer_range lowerCamelCase_ =num_labels lowerCamelCase_ =num_choices lowerCamelCase_ =summary_type lowerCamelCase_ =use_proj lowerCamelCase_ =scope def _snake_case ( self )-> Dict: lowerCamelCase_ =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ =random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase_ =None if self.use_input_lengths: lowerCamelCase_ =( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length lowerCamelCase_ =None if self.use_token_type_ids: lowerCamelCase_ =ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) lowerCamelCase_ =None lowerCamelCase_ =None lowerCamelCase_ =None if self.use_labels: lowerCamelCase_ =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ =ids_tensor([self.batch_size] , 2 ).float() lowerCamelCase_ =ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ =self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _snake_case ( self )-> List[str]: return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> str: lowerCamelCase_ =FlaubertModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , lengths=_SCREAMING_SNAKE_CASE , langs=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , langs=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> List[Any]: lowerCamelCase_ =FlaubertWithLMHeadModel(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> Optional[Any]: lowerCamelCase_ =FlaubertForQuestionAnsweringSimple(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> Optional[int]: lowerCamelCase_ =FlaubertForQuestionAnswering(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =model( _SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE , cls_index=_SCREAMING_SNAKE_CASE , is_impossible=_SCREAMING_SNAKE_CASE , p_mask=_SCREAMING_SNAKE_CASE , ) lowerCamelCase_ =model( _SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE , cls_index=_SCREAMING_SNAKE_CASE , is_impossible=_SCREAMING_SNAKE_CASE , ) ((lowerCamelCase_) , ) =result_with_labels.to_tuple() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE ) ((lowerCamelCase_) , ) =result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> Any: lowerCamelCase_ =FlaubertForSequenceClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> List[Any]: lowerCamelCase_ =self.num_labels lowerCamelCase_ =FlaubertForTokenClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , )-> Dict: lowerCamelCase_ =self.num_choices lowerCamelCase_ =FlaubertForMultipleChoice(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCamelCase_ =input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase_ =token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase_ =input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase_ =model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _snake_case ( self )-> int: lowerCamelCase_ =self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ) =config_and_inputs lowerCamelCase_ ={ """input_ids""": input_ids, """token_type_ids""": token_type_ids, """lengths""": input_lengths, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase): _UpperCamelCase:str = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) _UpperCamelCase:str = ( { "feature-extraction": FlaubertModel, "fill-mask": FlaubertWithLMHeadModel, "question-answering": FlaubertForQuestionAnsweringSimple, "text-classification": FlaubertForSequenceClassification, "token-classification": FlaubertForTokenClassification, "zero-shot": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )-> Optional[Any]: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False )-> List[Any]: lowerCamelCase_ =super()._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": lowerCamelCase_ =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) return inputs_dict def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ =FlaubertModelTester(self ) lowerCamelCase_ =ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , emb_dim=37 ) def _snake_case ( self )-> Optional[Any]: self.config_tester.run_common_tests() def _snake_case ( self )-> List[Any]: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> int: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> Tuple: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> List[Any]: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> List[Any]: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )-> List[str]: lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*_SCREAMING_SNAKE_CASE ) @slow def _snake_case ( self )-> Optional[Any]: for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ =FlaubertModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) @slow @require_torch_gpu def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ , lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return lowerCamelCase_ =True lowerCamelCase_ =model_class(config=_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCamelCase_ =torch.jit.trace( _SCREAMING_SNAKE_CASE , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_SCREAMING_SNAKE_CASE , os.path.join(_SCREAMING_SNAKE_CASE , """traced_model.pt""" ) ) lowerCamelCase_ =torch.jit.load(os.path.join(_SCREAMING_SNAKE_CASE , """traced_model.pt""" ) , map_location=_SCREAMING_SNAKE_CASE ) loaded(inputs_dict["""input_ids"""].to(_SCREAMING_SNAKE_CASE ) , inputs_dict["""attention_mask"""].to(_SCREAMING_SNAKE_CASE ) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase): @slow def _snake_case ( self )-> Union[str, Any]: lowerCamelCase_ =FlaubertModel.from_pretrained("""flaubert/flaubert_base_cased""" ) lowerCamelCase_ =torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) with torch.no_grad(): lowerCamelCase_ =model(_SCREAMING_SNAKE_CASE )[0] lowerCamelCase_ =torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE ) lowerCamelCase_ =torch.tensor( [[[-2.6_2_5_1, -1.4_2_9_8, -0.0_2_2_7], [-2.8_5_1_0, -1.6_3_8_7, 0.2_2_5_8], [-2.8_1_1_4, -1.1_8_3_2, -0.3_0_6_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
154
1
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def UpperCamelCase( lowercase_ ) -> List[Any]: '''simple docstring''' snake_case_ = [] embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', f'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', f'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', f'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', f'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def UpperCamelCase( lowercase_ , lowercase_ ) -> Optional[int]: '''simple docstring''' snake_case_ = [] attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', f'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', f'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', f'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', f'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def UpperCamelCase( lowercase_ ) -> Optional[Any]: '''simple docstring''' snake_case_ = [] token.append((f'''cvt.encoder.stages.{idx}.cls_token''', """stage2.cls_token""") ) return token def UpperCamelCase( ) -> Any: '''simple docstring''' snake_case_ = [] head.append(("""layernorm.weight""", """norm.weight""") ) head.append(("""layernorm.bias""", """norm.bias""") ) head.append(("""classifier.weight""", """head.weight""") ) head.append(("""classifier.bias""", """head.bias""") ) return head def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> List[Any]: '''simple docstring''' snake_case_ = """imagenet-1k-id2label.json""" snake_case_ = 1000 snake_case_ = """huggingface/label-files""" snake_case_ = num_labels snake_case_ = json.load(open(cached_download(hf_hub_url(lowercase_ , lowercase_ , repo_type="""dataset""" ) ) , """r""" ) ) snake_case_ = {int(lowercase_ ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} snake_case_ = snake_case_ = CvtConfig(num_labels=lowercase_ , idalabel=lowercase_ , labelaid=lowercase_ ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit("""/""" , 1 )[-1][4:6] == "13": snake_case_ = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit("""/""" , 1 )[-1][4:6] == "21": snake_case_ = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: snake_case_ = [2, 2, 20] snake_case_ = [3, 12, 16] snake_case_ = [192, 768, 1024] snake_case_ = CvtForImageClassification(lowercase_ ) snake_case_ = AutoImageProcessor.from_pretrained("""facebook/convnext-base-224-22k-1k""" ) snake_case_ = image_size snake_case_ = torch.load(lowercase_ , map_location=torch.device("""cpu""" ) ) snake_case_ = OrderedDict() snake_case_ = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: snake_case_ = list_of_state_dict + cls_token(lowercase_ ) snake_case_ = list_of_state_dict + embeddings(lowercase_ ) for cnt in range(config.depth[idx] ): snake_case_ = list_of_state_dict + attention(lowercase_ , lowercase_ ) snake_case_ = list_of_state_dict + final() for gg in list_of_state_dict: print(lowercase_ ) for i in range(len(lowercase_ ) ): snake_case_ = original_weights[list_of_state_dict[i][1]] model.load_state_dict(lowercase_ ) model.save_pretrained(lowercase_ ) image_processor.save_pretrained(lowercase_ ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( '''--cvt_model''', default='''cvt-w24''', type=str, help='''Name of the cvt model you\'d like to convert.''', ) parser.add_argument( '''--image_size''', default=384, type=int, help='''Input Image Size''', ) parser.add_argument( '''--cvt_file_name''', default=R'''cvtmodels\CvT-w24-384x384-IN-22k.pth''', type=str, help='''Input Image Size''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCamelCase_ = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
34
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''google/mobilenet_v1_1.0_224''': '''https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json''', '''google/mobilenet_v1_0.75_192''': '''https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : List[str] = 'mobilenet_v1' def __init__( self , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=1.0 , lowerCamelCase=8 , lowerCamelCase="relu6" , lowerCamelCase=True , lowerCamelCase=0.999 , lowerCamelCase=0.02 , lowerCamelCase=0.001 , **lowerCamelCase , ) -> List[str]: super().__init__(**lowerCamelCase ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) snake_case_ = num_channels snake_case_ = image_size snake_case_ = depth_multiplier snake_case_ = min_depth snake_case_ = hidden_act snake_case_ = tf_padding snake_case_ = classifier_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : str = version.parse('1.11' ) @property def lowerCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self ) -> float: return 1e-4
34
1
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch __snake_case = random.Random() def _A ( _lowercase , _lowercase=1.0 , _lowercase=None , _lowercase=None ) -> Tuple: """simple docstring""" if rng is None: __UpperCamelCase = global_rng __UpperCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __lowerCamelCase (unittest.TestCase ): def __init__( self: Optional[int],A_: Union[str, Any],A_: Optional[int]=7,A_: int=400,A_: Union[str, Any]=2000,A_: Union[str, Any]=1,A_: Tuple=0.0,A_: Any=1_6000,A_: List[Any]=True,A_: Any=80,A_: str=16,A_: List[str]=64,A_: Tuple="hann_window",A_: Optional[Any]=80,A_: Tuple=7600,A_: str=1E-10,A_: str=True,): '''simple docstring''' __UpperCamelCase = parent __UpperCamelCase = batch_size __UpperCamelCase = min_seq_length __UpperCamelCase = max_seq_length __UpperCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __UpperCamelCase = feature_size __UpperCamelCase = padding_value __UpperCamelCase = sampling_rate __UpperCamelCase = do_normalize __UpperCamelCase = num_mel_bins __UpperCamelCase = hop_length __UpperCamelCase = win_length __UpperCamelCase = win_function __UpperCamelCase = fmin __UpperCamelCase = fmax __UpperCamelCase = mel_floor __UpperCamelCase = return_attention_mask def snake_case_ ( self: List[str] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def snake_case_ ( self: Dict,A_: str=False,A_: Optional[Any]=False ): '''simple docstring''' def _flatten(A_: str ): return list(itertools.chain(*_A ) ) if equal_length: __UpperCamelCase = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __UpperCamelCase = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length,self.max_seq_length,self.seq_length_diff ) ] if numpify: __UpperCamelCase = [np.asarray(_A ) for x in speech_inputs] return speech_inputs def snake_case_ ( self: Optional[Any],A_: Optional[Any]=False,A_: Tuple=False ): '''simple docstring''' if equal_length: __UpperCamelCase = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __UpperCamelCase = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length,self.max_seq_length,self.seq_length_diff ) ] if numpify: __UpperCamelCase = [np.asarray(_A ) for x in speech_inputs] return speech_inputs @require_torch class __lowerCamelCase (a_ , unittest.TestCase ): _lowercase = SpeechTaFeatureExtractor def snake_case_ ( self: Optional[int] ): '''simple docstring''' __UpperCamelCase = SpeechTaFeatureExtractionTester(self ) def snake_case_ ( self: Optional[int],A_: Optional[int] ): '''simple docstring''' self.assertTrue(np.all(np.mean(_A,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_A,axis=0 ) - 1 ) < 1E-3 ) ) def snake_case_ ( self: List[str] ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = [np.asarray(_A ) for speech_input in speech_inputs] # Test not batched input __UpperCamelCase = feat_extract(speech_inputs[0],return_tensors='np' ).input_values __UpperCamelCase = feat_extract(np_speech_inputs[0],return_tensors='np' ).input_values self.assertTrue(np.allclose(_A,_A,atol=1E-3 ) ) # Test batched __UpperCamelCase = feat_extract(_A,return_tensors='np' ).input_values __UpperCamelCase = feat_extract(_A,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(_A,_A ): self.assertTrue(np.allclose(_A,_A,atol=1E-3 ) ) def snake_case_ ( self: Tuple ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = ['longest', 'max_length', 'do_not_pad'] __UpperCamelCase = [None, 1600, None] for max_length, padding in zip(_A,_A ): __UpperCamelCase = feat_extract(_A,padding=_A,max_length=_A,return_tensors='np' ) __UpperCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def snake_case_ ( self: Optional[int] ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = range(800,1400,200 ) __UpperCamelCase = [floats_list((1, x) )[0] for x in lengths] __UpperCamelCase = ['longest', 'max_length', 'do_not_pad'] __UpperCamelCase = [None, 1600, None] for max_length, padding in zip(_A,_A ): __UpperCamelCase = feat_extract(_A,max_length=_A,padding=_A ) __UpperCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def snake_case_ ( self: str ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = feat_extract( _A,truncation=_A,max_length=1000,padding='max_length',return_tensors='np' ) __UpperCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def snake_case_ ( self: int ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = feat_extract( _A,truncation=_A,max_length=1000,padding='longest',return_tensors='np' ) __UpperCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = feat_extract( _A,truncation=_A,max_length=2000,padding='longest',return_tensors='np' ) __UpperCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def snake_case_ ( self: Dict ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = np.random.rand(100 ).astype(np.floataa ) __UpperCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __UpperCamelCase = feature_extractor.pad([{'input_values': inputs}],return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __UpperCamelCase = feature_extractor.pad([{'input_values': inputs}],return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def snake_case_ ( self: Dict ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = [np.asarray(_A ) for speech_input in speech_inputs] # Test feature size __UpperCamelCase = feature_extractor(audio_target=_A,padding=_A,return_tensors='np' ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input __UpperCamelCase = feature_extractor(speech_inputs[0],return_tensors='np' ).input_values __UpperCamelCase = feature_extractor(np_speech_inputs[0],return_tensors='np' ).input_values self.assertTrue(np.allclose(_A,_A,atol=1E-3 ) ) # Test batched __UpperCamelCase = feature_extractor(_A,return_tensors='np' ).input_values __UpperCamelCase = feature_extractor(_A,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(_A,_A ): self.assertTrue(np.allclose(_A,_A,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __UpperCamelCase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __UpperCamelCase = np.asarray(_A ) __UpperCamelCase = feature_extractor(_A,return_tensors='np' ).input_values __UpperCamelCase = feature_extractor(_A,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(_A,_A ): self.assertTrue(np.allclose(_A,_A,atol=1E-3 ) ) def snake_case_ ( self: Union[str, Any] ): '''simple docstring''' __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __UpperCamelCase = feat_extract.model_input_names[0] __UpperCamelCase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_A ) == len(_A ) for x, y in zip(_A,processed_features[input_name] ) ) ) __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target(equal_length=_A ) __UpperCamelCase = BatchFeature({input_name: speech_inputs},tensor_type='np' ) __UpperCamelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __UpperCamelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def snake_case_ ( self: Optional[Any] ): '''simple docstring''' __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target(equal_length=_A ) __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __UpperCamelCase = feat_extract.model_input_names[0] __UpperCamelCase = BatchFeature({input_name: speech_inputs},tensor_type='pt' ) __UpperCamelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __UpperCamelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def snake_case_ ( self: Optional[Any] ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __UpperCamelCase = feat_extract.model_input_names[0] __UpperCamelCase = BatchFeature({input_name: speech_inputs} ) __UpperCamelCase = feat_extract.num_mel_bins # hack! __UpperCamelCase = feat_extract.pad(_A,padding='longest',return_tensors='np' )[input_name] __UpperCamelCase = feat_extract.pad(_A,padding='longest',return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def snake_case_ ( self: List[Any] ): '''simple docstring''' __UpperCamelCase = self.feat_extract_dict __UpperCamelCase = True __UpperCamelCase = self.feature_extraction_class(**_A ) __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __UpperCamelCase = [len(_A ) for x in speech_inputs] __UpperCamelCase = feat_extract.model_input_names[0] __UpperCamelCase = BatchFeature({input_name: speech_inputs} ) __UpperCamelCase = feat_extract.num_mel_bins # hack! __UpperCamelCase = feat_extract.pad(_A,padding='longest',return_tensors='np' ) self.assertIn('attention_mask',_A ) self.assertListEqual(list(processed.attention_mask.shape ),list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist(),_A ) def snake_case_ ( self: Optional[Any] ): '''simple docstring''' __UpperCamelCase = self.feat_extract_dict __UpperCamelCase = True __UpperCamelCase = self.feature_extraction_class(**_A ) __UpperCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __UpperCamelCase = [len(_A ) for x in speech_inputs] __UpperCamelCase = feat_extract.model_input_names[0] __UpperCamelCase = BatchFeature({input_name: speech_inputs} ) __UpperCamelCase = min(_A ) __UpperCamelCase = feat_extract.num_mel_bins # hack! __UpperCamelCase = feat_extract.pad( _A,padding='max_length',max_length=_A,truncation=_A,return_tensors='np' ) self.assertIn('attention_mask',_A ) self.assertListEqual( list(processed_pad.attention_mask.shape ),[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist(),[max_length for x in speech_inputs] ) def snake_case_ ( self: int,A_: List[str] ): '''simple docstring''' from datasets import load_dataset __UpperCamelCase = load_dataset('hf-internal-testing/librispeech_asr_dummy','clean',split='validation' ) # automatic decoding with librispeech __UpperCamelCase = ds.sort('id' ).select(range(_A ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def snake_case_ ( self: Union[str, Any] ): '''simple docstring''' __UpperCamelCase = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on __UpperCamelCase = self._load_datasamples(1 ) __UpperCamelCase = SpeechTaFeatureExtractor() __UpperCamelCase = feature_extractor(_A,return_tensors='pt' ).input_values self.assertEquals(input_values.shape,(1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30],_A,atol=1E-6 ) ) def snake_case_ ( self: str ): '''simple docstring''' __UpperCamelCase = torch.tensor( [-2.6_8_7_0, -3.0_1_0_4, -3.1_3_5_6, -3.5_3_5_2, -3.0_0_4_4, -3.0_3_5_3, -3.4_7_1_9, -3.6_7_7_7, -3.1_5_2_0, -2.9_4_3_5, -2.6_5_5_3, -2.8_7_9_5, -2.9_9_4_4, -2.5_9_2_1, -3.0_2_7_9, -3.0_3_8_6, -3.0_8_6_4, -3.1_2_9_1, -3.2_3_5_3, -2.7_4_4_4, -2.6_8_3_1, -2.7_2_8_7, -3.1_7_6_1, -3.1_5_7_1, -3.2_7_2_6, -3.0_5_8_2, -3.1_0_0_7, -3.4_5_3_3, -3.4_6_9_5, -3.0_9_9_8] ) # fmt: on __UpperCamelCase = self._load_datasamples(1 ) __UpperCamelCase = SpeechTaFeatureExtractor() __UpperCamelCase = feature_extractor(audio_target=_A,return_tensors='pt' ).input_values self.assertEquals(input_values.shape,(1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30],_A,atol=1E-4 ) )
310
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) __A : Union[str, Any] = { "configuration_speech_to_text": ["SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Speech2TextConfig"], "processing_speech_to_text": ["Speech2TextProcessor"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = ["Speech2TextTokenizer"] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = ["Speech2TextFeatureExtractor"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = [ "TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFSpeech2TextForConditionalGeneration", "TFSpeech2TextModel", "TFSpeech2TextPreTrainedModel", ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ "SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "Speech2TextForConditionalGeneration", "Speech2TextModel", "Speech2TextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys __A : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
273
0
lowerCAmelCase = 9.8_0_6_6_5 def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = g ): """simple docstring""" if fluid_density <= 0: raise ValueError('''Impossible fluid density''' ) if volume < 0: raise ValueError('''Impossible Object volume''' ) if gravity <= 0: raise ValueError('''Impossible Gravity''' ) return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
93
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase = { 'configuration_luke': ['LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LukeConfig'], 'tokenization_luke': ['LukeTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ 'LUKE_PRETRAINED_MODEL_ARCHIVE_LIST', 'LukeForEntityClassification', 'LukeForEntityPairClassification', 'LukeForEntitySpanClassification', 'LukeForMultipleChoice', 'LukeForQuestionAnswering', 'LukeForSequenceClassification', 'LukeForTokenClassification', 'LukeForMaskedLM', 'LukeModel', 'LukePreTrainedModel', ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
93
1
'''simple docstring''' from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def lowerCAmelCase_ ( _lowerCamelCase: NDArray[floataa] , _lowerCamelCase: NDArray[floataa] , _lowerCamelCase: list[int] , _lowerCamelCase: int , ): __SCREAMING_SNAKE_CASE : Dict = coefficient_matrix.shape __SCREAMING_SNAKE_CASE : List[Any] = constant_matrix.shape if rowsa != colsa: __SCREAMING_SNAKE_CASE : Dict = F"Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}" raise ValueError(_lowerCamelCase ) if colsa != 1: __SCREAMING_SNAKE_CASE : int = F"Constant matrix must be nx1 but received {rowsa}x{colsa}" raise ValueError(_lowerCamelCase ) if rowsa != rowsa: __SCREAMING_SNAKE_CASE : Optional[int] = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " F"received {rowsa}x{colsa} and {rowsa}x{colsa}" ) raise ValueError(_lowerCamelCase ) if len(_lowerCamelCase ) != rowsa: __SCREAMING_SNAKE_CASE : Tuple = ( "Number of initial values must be equal to number of rows in coefficient " F"matrix but received {len(_lowerCamelCase )} and {rowsa}" ) raise ValueError(_lowerCamelCase ) if iterations <= 0: raise ValueError("""Iterations must be at least 1""" ) __SCREAMING_SNAKE_CASE : NDArray[floataa] = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) __SCREAMING_SNAKE_CASE : Tuple = table.shape strictly_diagonally_dominant(_lowerCamelCase ) # Iterates the whole matrix for given number of times for _ in range(_lowerCamelCase ): __SCREAMING_SNAKE_CASE : Dict = [] for row in range(_lowerCamelCase ): __SCREAMING_SNAKE_CASE : Dict = 0 for col in range(_lowerCamelCase ): if col == row: __SCREAMING_SNAKE_CASE : Optional[Any] = table[row][col] elif col == cols - 1: __SCREAMING_SNAKE_CASE : Optional[Any] = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] __SCREAMING_SNAKE_CASE : Tuple = (temp + val) / denom new_val.append(_lowerCamelCase ) __SCREAMING_SNAKE_CASE : Optional[int] = new_val return [float(_lowerCamelCase ) for i in new_val] def lowerCAmelCase_ ( _lowerCamelCase: NDArray[floataa] ): __SCREAMING_SNAKE_CASE : Dict = table.shape __SCREAMING_SNAKE_CASE : int = True for i in range(0 , _lowerCamelCase ): __SCREAMING_SNAKE_CASE : Any = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("""Coefficient matrix is not strictly diagonally dominant""" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
112
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False')) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env') @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'pytorch', 'script': 'run_ddp.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'tensorflow', 'script': 'run_tf_dist.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.6, 'eval_loss': 0.7}, }, ]) class lowerCamelCase__ ( unittest.TestCase): '''simple docstring''' def _lowerCamelCase ( self :List[Any] ) -> Any: if self.framework == "pytorch": subprocess.run( f'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() , encoding="utf-8" , check=a , ) assert hasattr(self , "env" ) def _lowerCamelCase ( self :Any , a :Optional[Any] ) -> Dict: __UpperCamelCase : str = f'{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}' # distributed data settings __UpperCamelCase : Optional[int] = {"smdistributed": {"dataparallel": {"enabled": True}}} if self.script != "run_ddp.py" else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=a , instance_count=a , instance_type=self.instance_type , debugger_hook_config=a , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=a , py_version="py36" , ) def _lowerCamelCase ( self :Dict , a :Dict ) -> Optional[int]: TrainingJobAnalytics(a ).export_csv(f'{self.env.test_path}/{job_name}_metrics.csv' ) @parameterized.expand([(2,)] ) def _lowerCamelCase ( self :Dict , a :Tuple ) -> List[Any]: # create estimator __UpperCamelCase : int = self.create_estimator(a ) # run training estimator.fit() # result dataframe __UpperCamelCase : Optional[int] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __UpperCamelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) __UpperCamelCase : Tuple = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __UpperCamelCase : int = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f'{estimator.latest_training_job.name}.json' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , a )
232
0
'''simple docstring''' import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class lowercase_ : """simple docstring""" def __init__( self : Dict , __lowerCamelCase : str ): """simple docstring""" if isinstance(__lowerCamelCase , __lowerCamelCase ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden _SCREAMING_SNAKE_CASE = deepcopy(__lowerCamelCase ) elif os.path.exists(__lowerCamelCase ): with io.open(__lowerCamelCase , "r" , encoding="utf-8" ) as f: _SCREAMING_SNAKE_CASE = json.load(__lowerCamelCase ) else: try: _SCREAMING_SNAKE_CASE = baseaa.urlsafe_baadecode(__lowerCamelCase ).decode("utf-8" ) _SCREAMING_SNAKE_CASE = json.loads(__lowerCamelCase ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( F"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) _SCREAMING_SNAKE_CASE = config self.set_stage_and_offload() def lowerCAmelCase_ ( self : Any ): """simple docstring""" # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. _SCREAMING_SNAKE_CASE = self.get_value("zero_optimization.stage" , -1 ) # offload _SCREAMING_SNAKE_CASE = False if self.is_zeroa() or self.is_zeroa(): _SCREAMING_SNAKE_CASE = set(["cpu", "nvme"] ) _SCREAMING_SNAKE_CASE = set( [ self.get_value("zero_optimization.offload_optimizer.device" ), self.get_value("zero_optimization.offload_param.device" ), ] ) if len(offload_devices & offload_devices_valid ) > 0: _SCREAMING_SNAKE_CASE = True def lowerCAmelCase_ ( self : Tuple , __lowerCamelCase : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.config # find the config node of interest if it exists _SCREAMING_SNAKE_CASE = ds_key_long.split("." ) _SCREAMING_SNAKE_CASE = nodes.pop() for node in nodes: _SCREAMING_SNAKE_CASE = config.get(__lowerCamelCase ) if config is None: return None, ds_key return config, ds_key def lowerCAmelCase_ ( self : Any , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int=None ): """simple docstring""" _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.find_config_node(__lowerCamelCase ) if config is None: return default return config.get(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( self : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any]=False ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.config # find the config node of interest if it exists _SCREAMING_SNAKE_CASE = ds_key_long.split("." ) for node in nodes: _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = config.get(__lowerCamelCase ) if config is None: if must_exist: raise ValueError(F"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(__lowerCamelCase ) def lowerCAmelCase_ ( self : Optional[int] , __lowerCamelCase : List[Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.get_value(__lowerCamelCase ) return False if value is None else bool(__lowerCamelCase ) def lowerCAmelCase_ ( self : Dict , __lowerCamelCase : Any ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.get_value(__lowerCamelCase ) return False if value is None else not bool(__lowerCamelCase ) def lowerCAmelCase_ ( self : List[Any] ): """simple docstring""" return self._stage == 2 def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" return self._stage == 3 def lowerCAmelCase_ ( self : List[Any] ): """simple docstring""" return self._offload class lowercase_ : """simple docstring""" def __init__( self : List[Any] , __lowerCamelCase : Optional[int] ): """simple docstring""" _SCREAMING_SNAKE_CASE = engine def lowerCAmelCase_ ( self : List[Any] , __lowerCamelCase : Union[str, Any] , **__lowerCamelCase : Union[str, Any] ): """simple docstring""" # runs backpropagation and handles mixed precision self.engine.backward(__lowerCamelCase , **__lowerCamelCase ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class lowercase_ ( A ): """simple docstring""" def __init__( self : Dict , __lowerCamelCase : List[Any] ): """simple docstring""" super().__init__(__lowerCamelCase , device_placement=__lowerCamelCase , scaler=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = hasattr(self.optimizer , "overflow" ) def lowerCAmelCase_ ( self : Optional[int] , __lowerCamelCase : Dict=None ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def lowerCAmelCase_ ( self : Any ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" if self.__has_overflow__: return self.optimizer.overflow return False class lowercase_ ( A ): """simple docstring""" def __init__( self : Any , __lowerCamelCase : int , __lowerCamelCase : List[Any] ): """simple docstring""" super().__init__(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class lowercase_ : """simple docstring""" def __init__( self : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int]=0.0_0_1 , __lowerCamelCase : Dict=0 , **__lowerCamelCase : Dict ): """simple docstring""" _SCREAMING_SNAKE_CASE = params _SCREAMING_SNAKE_CASE = lr _SCREAMING_SNAKE_CASE = weight_decay _SCREAMING_SNAKE_CASE = kwargs class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=0 , **__lowerCamelCase : Dict ): """simple docstring""" _SCREAMING_SNAKE_CASE = optimizer _SCREAMING_SNAKE_CASE = total_num_steps _SCREAMING_SNAKE_CASE = warmup_num_steps _SCREAMING_SNAKE_CASE = kwargs
111
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ ( __A : Sequence[int] | None = None ) -> int: if nums is None or not nums: raise ValueError("Input sequence should not be empty" ) _SCREAMING_SNAKE_CASE = nums[0] for i in range(1 , len(__A ) ): _SCREAMING_SNAKE_CASE = nums[i] _SCREAMING_SNAKE_CASE = max(__A , ans + num , __A ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user lowerCamelCase_ = int(input('Enter number of elements : ').strip()) lowerCamelCase_ = list(map(int, input('\nEnter the numbers : ').strip().split()))[:n] print(max_subsequence_sum(array))
111
1
"""simple docstring""" import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging _A = logging.get_logger(__name__) logging.set_verbosity_info() def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: if "xprophetnet" in prophetnet_checkpoint_path: lowerCAmelCase__ : Any = XLMProphetNetForConditionalGenerationOld.from_pretrained(__UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ : str = XLMProphetNetForConditionalGeneration.from_pretrained( __UpperCAmelCase , output_loading_info=__UpperCAmelCase ) else: lowerCAmelCase__ : Tuple = ProphetNetForConditionalGenerationOld.from_pretrained(__UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ : List[str] = ProphetNetForConditionalGeneration.from_pretrained( __UpperCAmelCase , output_loading_info=__UpperCAmelCase ) lowerCAmelCase__ : int = ["""key_proj""", """value_proj""", """query_proj"""] lowerCAmelCase__ : Any = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: lowerCAmelCase__ : Optional[Any] = key.split(""".""" ) if attributes[0] == "lm_head": lowerCAmelCase__ : Optional[int] = prophet lowerCAmelCase__ : Dict = prophet_old else: lowerCAmelCase__ : Tuple = prophet.prophetnet lowerCAmelCase__ : Union[str, Any] = prophet_old.model lowerCAmelCase__ : str = False for attribute in attributes: if attribute in mapping: lowerCAmelCase__ : Optional[Any] = mapping[attribute] if not hasattr(__UpperCAmelCase , __UpperCAmelCase ) and len(__UpperCAmelCase ) > 0: lowerCAmelCase__ : Optional[int] = attribute elif hasattr(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ : int = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" lowerCAmelCase__ : List[str] = old_model.weight logger.info(f"""{attribute} is initialized.""" ) lowerCAmelCase__ : str = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" lowerCAmelCase__ : Dict = old_model.bias logger.info(f"""{attribute} is initialized""" ) lowerCAmelCase__ : int = True break elif attribute in special_keys and hasattr(__UpperCAmelCase , """in_proj_weight""" ): lowerCAmelCase__ : Optional[int] = old_model.in_proj_weight.shape[0] // 3 lowerCAmelCase__ : Dict = getattr(__UpperCAmelCase , __UpperCAmelCase ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": lowerCAmelCase__ : Optional[Any] = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) lowerCAmelCase__ : Optional[Any] = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": lowerCAmelCase__ : Optional[Any] = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) lowerCAmelCase__ : int = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": lowerCAmelCase__ : int = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) lowerCAmelCase__ : Dict = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) lowerCAmelCase__ : Dict = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." lowerCAmelCase__ : Optional[Any] = nn.Parameter(old_model.embed_positions.weight[:512, :] ) lowerCAmelCase__ : int = True break if attribute.isdigit(): lowerCAmelCase__ : List[str] = model[int(__UpperCAmelCase )] lowerCAmelCase__ : Any = old_model[int(__UpperCAmelCase )] else: lowerCAmelCase__ : Union[str, Any] = getattr(__UpperCAmelCase , __UpperCAmelCase ) if old_attribute == "": lowerCAmelCase__ : Optional[Any] = old_model else: if not hasattr(__UpperCAmelCase , __UpperCAmelCase ): raise ValueError(f"""{old_model} does not have {old_attribute}""" ) lowerCAmelCase__ : Tuple = getattr(__UpperCAmelCase , __UpperCAmelCase ) if not is_key_init: raise ValueError(f"""{key} was not correctly initialized!""" ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) prophet.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _A = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
242
"""simple docstring""" from __future__ import annotations import bisect def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 0 , __UpperCAmelCase = -1 ) -> int: if hi < 0: lowerCAmelCase__ : Union[str, Any] = len(__UpperCAmelCase ) while lo < hi: lowerCAmelCase__ : Tuple = lo + (hi - lo) // 2 if sorted_collection[mid] < item: lowerCAmelCase__ : Optional[int] = mid + 1 else: lowerCAmelCase__ : List[Any] = mid return lo def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 0 , __UpperCAmelCase = -1 ) -> int: if hi < 0: lowerCAmelCase__ : Union[str, Any] = len(__UpperCAmelCase ) while lo < hi: lowerCAmelCase__ : List[str] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: lowerCAmelCase__ : Dict = mid + 1 else: lowerCAmelCase__ : Any = mid return lo def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 0 , __UpperCAmelCase = -1 ) -> None: sorted_collection.insert(bisect_left(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 0 , __UpperCAmelCase = -1 ) -> None: sorted_collection.insert(bisect_right(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> int | None: lowerCAmelCase__ : Any = 0 lowerCAmelCase__ : Union[str, Any] = len(__UpperCAmelCase ) - 1 while left <= right: lowerCAmelCase__ : str = left + (right - left) // 2 lowerCAmelCase__ : List[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: lowerCAmelCase__ : Optional[int] = midpoint - 1 else: lowerCAmelCase__ : Optional[int] = midpoint + 1 return None def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> int | None: lowerCAmelCase__ : Any = bisect.bisect_left(__UpperCAmelCase , __UpperCAmelCase ) if index != len(__UpperCAmelCase ) and sorted_collection[index] == item: return index return None def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int | None: if right < left: return None lowerCAmelCase__ : List[str] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , midpoint - 1 ) else: return binary_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , midpoint + 1 , __UpperCAmelCase ) if __name__ == "__main__": _A = input("""Enter numbers separated by comma:\n""").strip() _A = sorted(int(item) for item in user_input.split(""",""")) _A = int(input("""Enter a single number to be found in the list:\n""")) _A = binary_search(collection, target) if result is None: print(f"""{target} was not found in {collection}.""") else: print(f"""{target} was found at position {result} in {collection}.""")
242
1
"""simple docstring""" import numpy as np def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : np.array ): return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
309
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase : Dict = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Tuple = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
309
1
"""simple docstring""" from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> None: if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): lowercase__, lowercase__: Union[str, Any] = array[indexa], array[indexa] def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> None: if length > 1: lowercase__: Union[str, Any] = int(length / 2 ) for i in range(__UpperCAmelCase , low + middle ): comp_and_swap(__UpperCAmelCase , __UpperCAmelCase , i + middle , __UpperCAmelCase ) bitonic_merge(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) bitonic_merge(__UpperCAmelCase , low + middle , __UpperCAmelCase , __UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> None: if length > 1: lowercase__: Optional[int] = int(length / 2 ) bitonic_sort(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , 1 ) bitonic_sort(__UpperCAmelCase , low + middle , __UpperCAmelCase , 0 ) bitonic_merge(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if __name__ == "__main__": __A = input("Enter numbers separated by a comma:\n").strip() __A = [int(item.strip()) for item in user_input.split(",")] bitonic_sort(unsorted, 0, len(unsorted), 1) print("\nSorted array in ascending order is: ", end="") print(*unsorted, sep=", ") bitonic_merge(unsorted, 0, len(unsorted), 0) print("Sorted array in descending order is: ", end="") print(*unsorted, sep=", ")
177
"""simple docstring""" from __future__ import annotations from math import pi, sqrt def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase ) -> tuple: if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
177
1
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _snake_case ( unittest.TestCase ): def snake_case__ ( self): UpperCAmelCase__ : List[str] = 10 def snake_case__ ( self): UpperCAmelCase__ : List[Any] = [1, 2, 3, 4] UpperCAmelCase__ : Optional[Any] = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(A__ , self.block_size , 0) , A__) def snake_case__ ( self): UpperCAmelCase__ : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] UpperCAmelCase__ : Optional[Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(A__ , self.block_size , 0) , A__) def snake_case__ ( self): UpperCAmelCase__ : int = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] UpperCAmelCase__ : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(A__ , self.block_size , 0) , A__) def snake_case__ ( self): UpperCAmelCase__ : Union[str, Any] = """It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.""" UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = process_story(A__) self.assertEqual(A__ , []) def snake_case__ ( self): UpperCAmelCase__ : Any = """""" UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = process_story(A__) self.assertEqual(A__ , []) self.assertEqual(A__ , []) def snake_case__ ( self): UpperCAmelCase__ : str = ( """It was the year of Our Lord one thousand seven hundred and """ """seventy-five\n\nSpiritual revelations were conceded to England """ """at that favoured period, as at this.\n@highlight\n\nIt was the best of times""" ) UpperCAmelCase__ , UpperCAmelCase__ : str = process_story(A__) UpperCAmelCase__ : Dict = [ """It was the year of Our Lord one thousand seven hundred and seventy-five.""", """Spiritual revelations were conceded to England at that favoured period, as at this.""", ] self.assertEqual(A__ , A__) UpperCAmelCase__ : List[Any] = ["""It was the best of times."""] self.assertEqual(A__ , A__) def snake_case__ ( self): UpperCAmelCase__ : int = torch.tensor([1, 2, 3, 4]) UpperCAmelCase__ : Dict = torch.tensor([1, 1, 1, 1]) np.testing.assert_array_equal(build_mask(A__ , 0).numpy() , expected.numpy()) def snake_case__ ( self): UpperCAmelCase__ : int = torch.tensor([1, 2, 3, 4, 23, 23, 23]) UpperCAmelCase__ : Union[str, Any] = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(A__ , 23).numpy() , expected.numpy()) def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = torch.tensor([8, 2, 3, 4, 1, 1, 1]) UpperCAmelCase__ : Optional[Any] = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(A__ , 1).numpy() , expected.numpy()) def snake_case__ ( self): UpperCAmelCase__ : List[Any] = 101 UpperCAmelCase__ : str = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]]) UpperCAmelCase__ : Optional[Any] = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]]) UpperCAmelCase__ : Tuple = compute_token_type_ids(A__ , A__) np.testing.assert_array_equal(A__ , A__)
359
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __A =logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__=False ): UpperCAmelCase__ : str = """backbone.""" if is_semantic else """""" UpperCAmelCase__ : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', """beit.embeddings.cls_token"""), (f'''{prefix}patch_embed.proj.weight''', """beit.embeddings.patch_embeddings.projection.weight"""), (f'''{prefix}patch_embed.proj.bias''', """beit.embeddings.patch_embeddings.projection.bias"""), (f'''{prefix}pos_embed''', """beit.embeddings.position_embeddings"""), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("""mask_token""", """beit.embeddings.mask_token"""), ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ] ) else: # layernorm + classification head rename_keys.extend( [ ("""fc_norm.weight""", """beit.pooler.layernorm.weight"""), ("""fc_norm.bias""", """beit.pooler.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__=False ): for i in range(config.num_hidden_layers ): UpperCAmelCase__ : Optional[Any] = """backbone.""" if is_semantic else """""" # queries, keys and values UpperCAmelCase__ : Any = state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase__ : List[str] = state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) UpperCAmelCase__ : int = state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) UpperCAmelCase__ : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase__ : Any = q_bias UpperCAmelCase__ : Dict = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase__ : Any = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase__ : Any = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained UpperCAmelCase__ : Dict = state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) UpperCAmelCase__ : Dict = state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) UpperCAmelCase__ : Union[str, Any] = gamma_a UpperCAmelCase__ : str = gamma_a def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : int = dct.pop(UpperCamelCase__ ) UpperCAmelCase__ : Optional[Any] = val def _UpperCamelCase ( ): UpperCAmelCase__ : Any = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase__ : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im @torch.no_grad() def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False ): UpperCAmelCase__ : Optional[Any] = False if """rvlcdip""" in checkpoint_url else True UpperCAmelCase__ : Any = BeitConfig(use_absolute_position_embeddings=UpperCamelCase__ , use_mask_token=UpperCamelCase__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: UpperCAmelCase__ : Optional[Any] = 1_0_2_4 UpperCAmelCase__ : Dict = 4_0_9_6 UpperCAmelCase__ : Any = 2_4 UpperCAmelCase__ : Tuple = 1_6 # labels if "rvlcdip" in checkpoint_url: UpperCAmelCase__ : int = 1_6 UpperCAmelCase__ : List[str] = """huggingface/label-files""" UpperCAmelCase__ : Optional[Any] = """rvlcdip-id2label.json""" UpperCAmelCase__ : Optional[int] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase__ : Union[str, Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} UpperCAmelCase__ : Optional[Any] = idalabel UpperCAmelCase__ : List[Any] = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys UpperCAmelCase__ : Optional[Any] = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" )["""model"""] UpperCAmelCase__ : List[str] = create_rename_keys(UpperCamelCase__ , has_lm_head=UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) read_in_q_k_v(UpperCamelCase__ , UpperCamelCase__ , has_lm_head=UpperCamelCase__ ) # load HuggingFace model UpperCAmelCase__ : str = BeitForMaskedImageModeling(UpperCamelCase__ ) if has_lm_head else BeitForImageClassification(UpperCamelCase__ ) model.eval() model.load_state_dict(UpperCamelCase__ ) # Check outputs on an image UpperCAmelCase__ : List[str] = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=UpperCamelCase__ ) UpperCAmelCase__ : List[str] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ) UpperCAmelCase__ : Optional[Any] = encoding["""pixel_values"""] UpperCAmelCase__ : Optional[Any] = model(UpperCamelCase__ ) UpperCAmelCase__ : int = outputs.logits # verify logits UpperCAmelCase__ : int = [1, 1_6] if """rvlcdip""" in checkpoint_url else [1, 1_9_6, 8_1_9_2] assert logits.shape == torch.Size(UpperCamelCase__ ), "Shape of logits not as expected" Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if push_to_hub: if has_lm_head: UpperCAmelCase__ : Union[str, Any] = """dit-base""" if """base""" in checkpoint_url else """dit-large""" else: UpperCAmelCase__ : Tuple = """dit-base-finetuned-rvlcdip""" if """dit-b""" in checkpoint_url else """dit-large-finetuned-rvlcdip""" image_processor.push_to_hub( repo_path_or_name=Path(UpperCamelCase__ , UpperCamelCase__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=UpperCamelCase__ , ) model.push_to_hub( repo_path_or_name=Path(UpperCamelCase__ , UpperCamelCase__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=UpperCamelCase__ , ) if __name__ == "__main__": __A =argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth', type=str, help='URL to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', ) __A =parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
283
0
import json import os import tempfile from unittest.mock import patch import torch from torch.utils.data import DataLoader, TensorDataset from accelerate import DistributedType, infer_auto_device_map, init_empty_weights from accelerate.accelerator import Accelerator from accelerate.state import GradientState, PartialState from accelerate.test_utils import require_bnb, require_multi_gpu, slow from accelerate.test_utils.testing import AccelerateTestCase, require_cuda from accelerate.utils import patch_environment def A_ ( ) -> int: a__ : Dict = torch.nn.Linear(2 , 4 ) a__ : Dict = torch.optim.AdamW(model.parameters() , lr=1.0 ) a__ : Any = torch.optim.lr_scheduler.OneCycleLR(A__ , max_lr=0.01 , steps_per_epoch=2 , epochs=1 ) a__ : Tuple = DataLoader(TensorDataset(torch.tensor([1, 2, 3] ) ) ) a__ : int = DataLoader(TensorDataset(torch.tensor([4, 5, 6] ) ) ) return model, optimizer, scheduler, train_dl, valid_dl def A_ ( A__ ) -> Optional[Any]: return (model.weight.abs().sum() + model.bias.abs().sum()).item() def A_ ( A__ ) -> str: a__ : List[Any] = torch.nn.Linear(*tuple(model.weight.T.shape ) ).state_dict() model.load_state_dict(A__ ) class A__ ( __UpperCAmelCase ): """simple docstring""" @require_cuda def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[str] = Accelerator() assert PartialState._shared_state["_cpu"] is False assert PartialState._shared_state["device"].type == "cuda" with self.assertRaises(lowercase): a__ : Union[str, Any] = Accelerator(cpu=lowercase) def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : Tuple = Accelerator() a__ : Optional[Any] = GradientState() assert state.num_steps == 1 a__ : Union[str, Any] = 4 assert state.num_steps == 4 assert state.sync_gradients is True a__ : Optional[int] = False assert state.sync_gradients is False GradientState._reset_state() def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : Tuple = Accelerator() a__ , a__ , a__ , a__ , a__ : Tuple = create_components() ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) : List[Any] = accelerator.prepare(lowercase , lowercase , lowercase , lowercase , lowercase) self.assertTrue(prepared_model in accelerator._models) self.assertTrue(prepared_optimizer in accelerator._optimizers) self.assertTrue(prepared_scheduler in accelerator._schedulers) self.assertTrue(prepared_train_dl in accelerator._dataloaders) self.assertTrue(prepared_valid_dl in accelerator._dataloaders) def __lowercase ( self) -> Dict: '''simple docstring''' a__ : Union[str, Any] = Accelerator() a__ , a__ , a__ , a__ , a__ : List[str] = create_components() accelerator.prepare(lowercase , lowercase , lowercase , lowercase , lowercase) accelerator.free_memory() self.assertTrue(len(accelerator._models) == 0) self.assertTrue(len(accelerator._optimizers) == 0) self.assertTrue(len(accelerator._schedulers) == 0) self.assertTrue(len(accelerator._dataloaders) == 0) def __lowercase ( self) -> str: '''simple docstring''' PartialState._reset_state() # Mock torch.cuda.set_device to avoid an exception as the device doesn't exist def noop(*lowercase , **lowercase): pass with patch('torch.cuda.set_device' , lowercase), patch_environment(ACCELERATE_TORCH_DEVICE='cuda:64'): a__ : Tuple = Accelerator() self.assertEqual(str(accelerator.state.device) , 'cuda:64') def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : Optional[int] = Accelerator() a__ , a__ , a__ , a__ , a__ : Union[str, Any] = create_components() accelerator.prepare(lowercase , lowercase , lowercase , lowercase , lowercase) a__ : Tuple = get_signature(lowercase) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(lowercase) # make sure random weights don't match load_random_weights(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) > 1e-3) # make sure loaded weights match accelerator.load_state(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) < 1e-3) def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : List[str] = Accelerator() a__ , a__ , a__ , a__ , a__ : Optional[int] = create_components() accelerator.prepare(lowercase , lowercase , lowercase , lowercase , lowercase) a__ : List[str] = get_signature(lowercase) # saving hook def save_config(lowercase , lowercase , lowercase): a__ : int = {'class_name': models[0].__class__.__name__} with open(os.path.join(lowercase , 'data.json') , 'w') as f: json.dump(lowercase , lowercase) # loading hook def load_config(lowercase , lowercase): with open(os.path.join(lowercase , 'data.json') , 'r') as f: a__ : List[Any] = json.load(lowercase) a__ : Union[str, Any] = config['class_name'] a__ : Optional[Any] = accelerator.register_save_state_pre_hook(lowercase) a__ : str = accelerator.register_load_state_pre_hook(lowercase) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(lowercase) # make sure random weights don't match with hooks load_random_weights(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) > 1e-3) # random class name to verify correct one is loaded a__ : int = 'random' # make sure loaded weights match with hooks accelerator.load_state(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) < 1e-3) # mode.class_name is loaded from config self.assertTrue(model.class_name == model.__class__.__name__) # remove hooks save_hook.remove() load_hook.remove() with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(lowercase) # make sure random weights don't match with hooks removed load_random_weights(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) > 1e-3) # random class name to verify correct one is loaded a__ : Union[str, Any] = 'random' # make sure loaded weights match with hooks removed accelerator.load_state(lowercase) self.assertTrue(abs(model_signature - get_signature(lowercase)) < 1e-3) # mode.class_name is NOT loaded from config self.assertTrue(model.class_name != model.__class__.__name__) def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : str = Accelerator() a__ , a__ , a__ , a__ , a__ : str = create_components() a__ : Dict = None # This should work a__ , a__ , a__ , a__ , a__ , a__ : Tuple = accelerator.prepare( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) self.assertTrue(dummy_obj is None) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : List[Any] = Accelerator() a__ , a__ , a__ , a__ , a__ : int = create_components() a__ : int = [1, 2, 3] # This should work a__ , a__ , a__ , a__ , a__ , a__ : Any = accelerator.prepare( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Dummy object should have `_is_accelerate_prepared` set to `True`' , ) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Model is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Optimizer is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Scheduler is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Train Dataloader is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(lowercase , '_is_accelerate_prepared' , lowercase) , lowercase , 'Valid Dataloader is missing `_is_accelerator_prepared` or is set to `False`' , ) @slow @require_bnb def __lowercase ( self) -> Optional[int]: '''simple docstring''' from transformers import AutoModelForCausalLM a__ : List[str] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=lowercase , device_map={'': 0} , ) a__ : Union[str, Any] = Accelerator() # This should work a__ : Union[str, Any] = accelerator.prepare(lowercase) @slow @require_bnb def __lowercase ( self) -> Any: '''simple docstring''' from transformers import AutoModelForCausalLM a__ : Dict = Accelerator() with init_empty_weights(): a__ : Dict = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) model.tie_weights() a__ : Optional[int] = infer_auto_device_map(lowercase) a__ : Optional[Any] = 'cpu' a__ : str = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , device_map=lowercase , load_in_abit=lowercase , llm_inta_enable_fpaa_cpu_offload=lowercase) # This should not work and get value error with self.assertRaises(lowercase): a__ : List[str] = accelerator.prepare(lowercase) @slow @require_bnb @require_multi_gpu def __lowercase ( self) -> str: '''simple docstring''' from transformers import AutoModelForCausalLM a__ : int = {'distributed_type': DistributedType.MULTI_GPU} with init_empty_weights(): a__ : Optional[int] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) model.tie_weights() a__ : Optional[int] = infer_auto_device_map(lowercase) a__ : Optional[Any] = 1 a__ : List[str] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=lowercase , device_map=lowercase , ) a__ : Optional[Any] = Accelerator() # This should not work and get value error with self.assertRaises(lowercase): a__ : int = accelerator.prepare(lowercase) PartialState._reset_state() @slow @require_bnb @require_multi_gpu def __lowercase ( self) -> Optional[int]: '''simple docstring''' from transformers import AutoModelForCausalLM with init_empty_weights(): a__ : int = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) a__ : List[str] = infer_auto_device_map(lowercase) a__ : List[str] = 1 a__ : List[str] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=lowercase , device_map=lowercase , ) a__ : int = Accelerator() # This should work a__ : Optional[Any] = accelerator.prepare(lowercase) @require_cuda def __lowercase ( self) -> Dict: '''simple docstring''' a__ : Optional[int] = torch.nn.Linear(10 , 10) a__ : Optional[int] = torch.optim.SGD(model.parameters() , lr=0.01) a__ : Optional[int] = Accelerator(cpu=lowercase) a__ : str = accelerator.prepare(lowercase)
99
def A_ ( A__ ) -> int: stooge(A__ , 0 , len(A__ ) - 1 ) return arr def A_ ( A__ , A__ , A__ ) -> List[Any]: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: a__ , a__ : List[str] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: a__ : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(A__ , A__ , (h - t) ) # Recursively sort last 2/3 elements stooge(A__ , i + t , (A__) ) # Recursively sort first 2/3 elements stooge(A__ , A__ , (h - t) ) if __name__ == "__main__": lowercase : Dict = input("""Enter numbers separated by a comma:\n""").strip() lowercase : Dict = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
99
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : '''simple docstring''' def __init__( self: Optional[int] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Optional[Any]=3 ,lowerCamelCase_: Tuple=32 ,lowerCamelCase_: Any=3 ,lowerCamelCase_: Optional[int]=10 ,lowerCamelCase_: Any=[10, 20, 30, 40] ,lowerCamelCase_: Union[str, Any]=[1, 1, 2, 1] ,lowerCamelCase_: Tuple=True ,lowerCamelCase_: str=True ,lowerCamelCase_: Union[str, Any]="relu" ,lowerCamelCase_: Optional[int]=3 ,lowerCamelCase_: Optional[int]=None ,) -> Union[str, Any]: UpperCAmelCase_ : List[Any] = parent UpperCAmelCase_ : Optional[int] = batch_size UpperCAmelCase_ : Optional[int] = image_size UpperCAmelCase_ : List[str] = num_channels UpperCAmelCase_ : Dict = embeddings_size UpperCAmelCase_ : Optional[int] = hidden_sizes UpperCAmelCase_ : Optional[Any] = depths UpperCAmelCase_ : Tuple = is_training UpperCAmelCase_ : Union[str, Any] = use_labels UpperCAmelCase_ : int = hidden_act UpperCAmelCase_ : Dict = num_labels UpperCAmelCase_ : Optional[Any] = scope UpperCAmelCase_ : Optional[Any] = len(lowerCamelCase_ ) def A__ ( self: str ) -> Any: UpperCAmelCase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : Union[str, Any] = None if self.use_labels: UpperCAmelCase_ : Tuple = ids_tensor([self.batch_size] ,self.num_labels ) UpperCAmelCase_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def A__ ( self: Tuple ) -> Union[str, Any]: return ResNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def A__ ( self: Optional[int] ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Dict ,lowerCamelCase_: Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : Any = TFResNetModel(config=lowerCamelCase_ ) UpperCAmelCase_ : Tuple = model(lowerCamelCase_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def A__ ( self: List[str] ,lowerCamelCase_: Any ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Any ) -> str: UpperCAmelCase_ : Optional[int] = self.num_labels UpperCAmelCase_ : Dict = TFResNetForImageClassification(lowerCamelCase_ ) UpperCAmelCase_ : int = model(lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def A__ ( self: Dict ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[str] = config_and_inputs UpperCAmelCase_ : Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Dict = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () A__ : Dict = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) A__ : Dict = False A__ : List[str] = False A__ : str = False A__ : Optional[int] = False A__ : int = False def A__ ( self: int ) -> Union[str, Any]: UpperCAmelCase_ : int = TFResNetModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self ,config_class=lowerCamelCase_ ,has_text_modality=lowerCamelCase_ ) def A__ ( self: List[Any] ) -> Union[str, Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self: Any ) -> Union[str, Any]: return @unittest.skip(reason="""ResNet does not use inputs_embeds""" ) def A__ ( self: Any ) -> Tuple: pass @unittest.skip(reason="""ResNet does not support input and output embeddings""" ) def A__ ( self: Optional[int] ) -> int: pass def A__ ( self: List[Any] ) -> List[str]: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Any = model_class(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : List[Any] = [*signature.parameters.keys()] UpperCAmelCase_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase_ ) def A__ ( self: List[Any] ) -> Dict: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def A__ ( self: str ) -> int: def check_hidden_states_output(lowerCamelCase_: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Any ): UpperCAmelCase_ : Any = model_class(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : Any = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase_ ) ,expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 4, self.model_tester.image_size // 4] ,) UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Any = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase_ : Tuple = layer_type UpperCAmelCase_ : List[str] = True check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : Dict = True check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) def A__ ( self: str ) -> Tuple: UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def A__ ( self: Union[str, Any] ) -> int: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Dict = TFResNetModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def lowerCamelCase_ ( ): '''simple docstring''' UpperCAmelCase_ : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _snake_case ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self: int ) -> List[str]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A__ ( self: Optional[Any] ) -> List[Any]: UpperCAmelCase_ : Optional[Any] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase_ : Tuple = self.default_image_processor UpperCAmelCase_ : List[str] = prepare_img() UpperCAmelCase_ : Union[str, Any] = image_processor(images=lowerCamelCase_ ,return_tensors="""tf""" ) # forward pass UpperCAmelCase_ : Dict = model(**lowerCamelCase_ ) # verify the logits UpperCAmelCase_ : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() ,lowerCamelCase_ ,atol=1e-4 ) )
59
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _snake_case ( __snake_case ): '''simple docstring''' A__ : Optional[int] = "roformer" def __init__( self: Optional[int] ,lowerCamelCase_: Tuple=50000 ,lowerCamelCase_: Optional[int]=None ,lowerCamelCase_: List[Any]=768 ,lowerCamelCase_: List[Any]=12 ,lowerCamelCase_: Optional[int]=12 ,lowerCamelCase_: Optional[Any]=3072 ,lowerCamelCase_: int="gelu" ,lowerCamelCase_: str=0.1 ,lowerCamelCase_: Union[str, Any]=0.1 ,lowerCamelCase_: Any=1536 ,lowerCamelCase_: str=2 ,lowerCamelCase_: Optional[int]=0.0_2 ,lowerCamelCase_: int=1e-12 ,lowerCamelCase_: Optional[int]=0 ,lowerCamelCase_: Any=False ,lowerCamelCase_: Union[str, Any]=True ,**lowerCamelCase_: List[str] ,) -> Tuple: super().__init__(pad_token_id=lowerCamelCase_ ,**lowerCamelCase_ ) UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Optional[int] = hidden_size if embedding_size is None else embedding_size UpperCAmelCase_ : Optional[Any] = hidden_size UpperCAmelCase_ : Optional[int] = num_hidden_layers UpperCAmelCase_ : Any = num_attention_heads UpperCAmelCase_ : Optional[Any] = hidden_act UpperCAmelCase_ : Any = intermediate_size UpperCAmelCase_ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase_ : Optional[int] = attention_probs_dropout_prob UpperCAmelCase_ : Tuple = max_position_embeddings UpperCAmelCase_ : Optional[Any] = type_vocab_size UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : Optional[Any] = rotary_value UpperCAmelCase_ : str = use_cache class _snake_case ( __snake_case ): '''simple docstring''' @property def A__ ( self: Tuple ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase_ : Tuple = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase_ : Optional[Any] = {0: """batch""", 1: """sequence"""} UpperCAmelCase_ : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
59
1
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self ) -> Tuple: """simple docstring""" super().__init__() UpperCamelCase__ : List[Any] = nn.Linear(3, 4 ) UpperCamelCase__ : Dict = nn.BatchNormad(4 ) UpperCamelCase__ : List[Any] = nn.Linear(4, 5 ) def UpperCamelCase__ ( self, __magic_name__ ) -> Optional[int]: """simple docstring""" return self.lineara(self.batchnorm(self.lineara(__magic_name__ ) ) ) class lowercase__ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ) -> Any: """simple docstring""" UpperCamelCase__ : str = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(__magic_name__, model.state_dict() ) UpperCamelCase__ : Any = os.path.join(__magic_name__, '''index.json''' ) self.assertTrue(os.path.isfile(__magic_name__ ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: UpperCamelCase__ : List[Any] = os.path.join(__magic_name__, f"{key}.dat" ) self.assertTrue(os.path.isfile(__magic_name__ ) ) # TODO: add tests on the fact weights are properly loaded def UpperCamelCase__ ( self ) -> str: """simple docstring""" UpperCamelCase__ : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: UpperCamelCase__ : List[Any] = torch.randn(2, 3, dtype=__magic_name__ ) with TemporaryDirectory() as tmp_dir: UpperCamelCase__ : Tuple = offload_weight(__magic_name__, '''weight''', __magic_name__, {} ) UpperCamelCase__ : Optional[Any] = os.path.join(__magic_name__, '''weight.dat''' ) self.assertTrue(os.path.isfile(__magic_name__ ) ) self.assertDictEqual(__magic_name__, {'''weight''': {'''shape''': [2, 3], '''dtype''': str(__magic_name__ ).split('''.''' )[1]}} ) UpperCamelCase__ : Optional[int] = load_offloaded_weight(__magic_name__, index['''weight'''] ) self.assertTrue(torch.equal(__magic_name__, __magic_name__ ) ) def UpperCamelCase__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase__ : Union[str, Any] = ModelForTest() UpperCamelCase__ : Optional[int] = model.state_dict() UpperCamelCase__ : List[str] = {k: v for k, v in state_dict.items() if '''linear2''' not in k} UpperCamelCase__ : int = {k: v for k, v in state_dict.items() if '''linear2''' in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__magic_name__, __magic_name__ ) UpperCamelCase__ : List[Any] = OffloadedWeightsLoader(state_dict=__magic_name__, save_folder=__magic_name__ ) # Every key is there with the right value self.assertEqual(sorted(__magic_name__ ), sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__magic_name__, weight_map[key] ) ) UpperCamelCase__ : Dict = {k: v for k, v in state_dict.items() if '''weight''' in k} UpperCamelCase__ : int = {k: v for k, v in state_dict.items() if '''weight''' not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__magic_name__, __magic_name__ ) UpperCamelCase__ : List[Any] = OffloadedWeightsLoader(state_dict=__magic_name__, save_folder=__magic_name__ ) # Every key is there with the right value self.assertEqual(sorted(__magic_name__ ), sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__magic_name__, weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(__magic_name__, __magic_name__ ) # Duplicates are removed UpperCamelCase__ : List[Any] = OffloadedWeightsLoader(state_dict=__magic_name__, save_folder=__magic_name__ ) # Every key is there with the right value self.assertEqual(sorted(__magic_name__ ), sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__magic_name__, weight_map[key] ) ) def UpperCamelCase__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase__ : Tuple = {'''a.1''': 0, '''a.10''': 1, '''a.2''': 2} UpperCamelCase__ : Tuple = extract_submodules_state_dict(__magic_name__, ['''a.1''', '''a.2'''] ) self.assertDictEqual(__magic_name__, {'''a.1''': 0, '''a.2''': 2} ) UpperCamelCase__ : Optional[int] = {'''a.1.a''': 0, '''a.10.a''': 1, '''a.2.a''': 2} UpperCamelCase__ : Optional[Any] = extract_submodules_state_dict(__magic_name__, ['''a.1''', '''a.2'''] ) self.assertDictEqual(__magic_name__, {'''a.1.a''': 0, '''a.2.a''': 2} )
201
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'google/mobilenet_v2_1.4_224': 'https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json', 'google/mobilenet_v2_1.0_224': 'https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json', 'google/mobilenet_v2_0.75_160': 'https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json', 'google/mobilenet_v2_0.35_96': 'https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : List[str] = "mobilenet_v2" def __init__( self, __magic_name__=3, __magic_name__=224, __magic_name__=1.0, __magic_name__=8, __magic_name__=8, __magic_name__=6, __magic_name__=32, __magic_name__=True, __magic_name__=True, __magic_name__="relu6", __magic_name__=True, __magic_name__=0.8, __magic_name__=0.02, __magic_name__=0.001, __magic_name__=255, **__magic_name__, ) -> List[Any]: """simple docstring""" super().__init__(**__magic_name__ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) UpperCamelCase__ : Union[str, Any] = num_channels UpperCamelCase__ : int = image_size UpperCamelCase__ : int = depth_multiplier UpperCamelCase__ : Tuple = depth_divisible_by UpperCamelCase__ : List[str] = min_depth UpperCamelCase__ : Optional[int] = expand_ratio UpperCamelCase__ : Optional[int] = output_stride UpperCamelCase__ : Tuple = first_layer_is_expansion UpperCamelCase__ : Union[str, Any] = finegrained_output UpperCamelCase__ : str = hidden_act UpperCamelCase__ : Optional[Any] = tf_padding UpperCamelCase__ : Optional[int] = classifier_dropout_prob UpperCamelCase__ : int = initializer_range UpperCamelCase__ : Union[str, Any] = layer_norm_eps UpperCamelCase__ : Tuple = semantic_loss_ignore_index class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : Union[str, Any] = version.parse("1.11" ) @property def UpperCamelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def UpperCamelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def UpperCamelCase__ ( self ) -> float: """simple docstring""" return 1E-4
201
1
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class lowercase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self, __magic_name__, __magic_name__ = True, __magic_name__ = None, __magic_name__ = 32, __magic_name__ = True, __magic_name__ = 1 / 255, __magic_name__ = True, __magic_name__ = True, __magic_name__ = [0.4814_5466, 0.457_8275, 0.4082_1073], __magic_name__ = [0.2686_2954, 0.2613_0258, 0.2757_7711], __magic_name__ = True, __magic_name__=7, __magic_name__=30, __magic_name__=400, __magic_name__=3, ) -> Optional[int]: """simple docstring""" UpperCamelCase__ : List[str] = parent UpperCamelCase__ : Optional[Any] = do_resize UpperCamelCase__ : Dict = size if size is not None else {'''shortest_edge''': 288} UpperCamelCase__ : Any = size_divisor UpperCamelCase__ : Dict = do_rescale UpperCamelCase__ : List[str] = rescale_factor UpperCamelCase__ : Tuple = do_normalize UpperCamelCase__ : Tuple = do_center_crop UpperCamelCase__ : Tuple = image_mean UpperCamelCase__ : Optional[int] = image_std UpperCamelCase__ : Tuple = do_pad UpperCamelCase__ : Dict = batch_size UpperCamelCase__ : Union[str, Any] = num_channels UpperCamelCase__ : Dict = min_resolution UpperCamelCase__ : Optional[int] = max_resolution def UpperCamelCase__ ( self ) -> str: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase__ ( self, __magic_name__, __magic_name__=False ) -> str: """simple docstring""" if not batched: UpperCamelCase__ : Dict = self.size['''shortest_edge'''] UpperCamelCase__ : Tuple = image_inputs[0] if isinstance(__magic_name__, Image.Image ): UpperCamelCase__ ,UpperCamelCase__ : Tuple = image.size else: UpperCamelCase__ ,UpperCamelCase__ : str = image.shape[1], image.shape[2] UpperCamelCase__ : Any = size / min(__magic_name__, __magic_name__ ) if h < w: UpperCamelCase__ ,UpperCamelCase__ : List[str] = size, scale * w else: UpperCamelCase__ ,UpperCamelCase__ : Union[str, Any] = scale * h, size UpperCamelCase__ : int = int((1333 / 800) * size ) if max(__magic_name__, __magic_name__ ) > max_size: UpperCamelCase__ : List[Any] = max_size / max(__magic_name__, __magic_name__ ) UpperCamelCase__ : Optional[Any] = newh * scale UpperCamelCase__ : int = neww * scale UpperCamelCase__ ,UpperCamelCase__ : Optional[Any] = int(newh + 0.5 ), int(neww + 0.5 ) UpperCamelCase__ ,UpperCamelCase__ : List[Any] = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCamelCase__ : Tuple = [] for image in image_inputs: UpperCamelCase__ ,UpperCamelCase__ : str = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase__ : Union[str, Any] = max(__magic_name__, key=lambda __magic_name__ : item[0] )[0] UpperCamelCase__ : Union[str, Any] = max(__magic_name__, key=lambda __magic_name__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowercase__ ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' a : Tuple = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase__ : Optional[int] = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase__ ( self ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) -> str: """simple docstring""" UpperCamelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__magic_name__, '''image_mean''' ) ) self.assertTrue(hasattr(__magic_name__, '''image_std''' ) ) self.assertTrue(hasattr(__magic_name__, '''do_normalize''' ) ) self.assertTrue(hasattr(__magic_name__, '''do_resize''' ) ) self.assertTrue(hasattr(__magic_name__, '''size''' ) ) self.assertTrue(hasattr(__magic_name__, '''size_divisor''' ) ) def UpperCamelCase__ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase__ ( self ) -> List[str]: """simple docstring""" # Initialize image processor UpperCamelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester, equal_resolution=__magic_name__ ) for image in image_inputs: self.assertIsInstance(__magic_name__, Image.Image ) # Test not batched input UpperCamelCase__ : Optional[Any] = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : Optional[Any] = self.image_processor_tester.get_expected_values(__magic_name__ ) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched UpperCamelCase__ : int = image_processing(__magic_name__, return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : Any = self.image_processor_tester.get_expected_values(__magic_name__, batched=__magic_name__ ) self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), ) def UpperCamelCase__ ( self ) -> List[str]: """simple docstring""" # Initialize image processor UpperCamelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase__ : Tuple = prepare_image_inputs(self.image_processor_tester, equal_resolution=__magic_name__, numpify=__magic_name__ ) for image in image_inputs: self.assertIsInstance(__magic_name__, np.ndarray ) # Test not batched input UpperCamelCase__ : List[str] = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : Tuple = self.image_processor_tester.get_expected_values(__magic_name__ ) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched UpperCamelCase__ : Union[str, Any] = image_processing(__magic_name__, return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : List[Any] = self.image_processor_tester.get_expected_values(__magic_name__, batched=__magic_name__ ) self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), ) def UpperCamelCase__ ( self ) -> Dict: """simple docstring""" # Initialize image processor UpperCamelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester, equal_resolution=__magic_name__, torchify=__magic_name__ ) for image in image_inputs: self.assertIsInstance(__magic_name__, torch.Tensor ) # Test not batched input UpperCamelCase__ : Any = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : Dict = self.image_processor_tester.get_expected_values(__magic_name__ ) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched UpperCamelCase__ : int = image_processing(__magic_name__, return_tensors='''pt''' ).pixel_values UpperCamelCase__ ,UpperCamelCase__ : Union[str, Any] = self.image_processor_tester.get_expected_values(__magic_name__, batched=__magic_name__ ) self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), )
247
import torch def lowerCAmelCase_ ( ) -> int: if torch.cuda.is_available(): UpperCamelCase__ : Optional[int] = torch.cuda.device_count() else: UpperCamelCase__ : int = 0 print(f"Successfully ran on {num_gpus} GPUs" ) if __name__ == "__main__": main()
247
1
"""simple docstring""" import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class SCREAMING_SNAKE_CASE__ ( snake_case__ , unittest.TestCase ): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @property def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' UpperCAmelCase : str = ort.SessionOptions() UpperCAmelCase : int = False return options def SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) UpperCAmelCase : Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) UpperCAmelCase : List[Any] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : Any = "A red cat sitting on a park bench" UpperCAmelCase : Optional[Any] = np.random.RandomState(0 ) UpperCAmelCase : Tuple = pipe( prompt=snake_case__ , image=snake_case__ , mask_image=snake_case__ , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case__ , output_type="""np""" , ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : Optional[int] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) UpperCAmelCase : Optional[Any] = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) UpperCAmelCase : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) UpperCAmelCase : Tuple = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-inpainting""" , subfolder="""scheduler""" , revision="""onnx""" ) UpperCAmelCase : Optional[int] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , scheduler=snake_case__ , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : Optional[int] = "A red cat sitting on a park bench" UpperCAmelCase : Tuple = np.random.RandomState(0 ) UpperCAmelCase : List[Any] = pipe( prompt=snake_case__ , image=snake_case__ , mask_image=snake_case__ , guidance_scale=7.5 , num_inference_steps=20 , generator=snake_case__ , output_type="""np""" , ) UpperCAmelCase : Tuple = output.images UpperCAmelCase : Tuple = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) UpperCAmelCase : Optional[int] = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
109
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class lowercase__ ( snake_case__ ): _UpperCAmelCase :BigBirdConfig _UpperCAmelCase :jnp.dtype = jnp.floataa _UpperCAmelCase :bool = True def UpperCAmelCase__ ( self : Dict ): super().setup() lowerCamelCase_ : List[str] =nn.Dense(5 , dtype=self.dtype ) def __call__( self : Dict , *snake_case__ : Optional[int] , **snake_case__ : Any ): lowerCamelCase_ : int =super().__call__(*snake_case__ , **snake_case__ ) lowerCamelCase_ : Tuple =self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class lowercase__ ( snake_case__ ): _UpperCAmelCase :List[str] = FlaxBigBirdForNaturalQuestionsModule def _snake_case ( lowerCamelCase__ : Tuple , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : int ) -> List[str]: def cross_entropy(lowerCamelCase__ : int , lowerCamelCase__ : List[Any] , lowerCamelCase__ : int=None ): lowerCamelCase_ : List[str] =logits.shape[-1] lowerCamelCase_ : List[str] =(labels[..., None] == jnp.arange(lowerCamelCase__ )[None]).astype("f4" ) lowerCamelCase_ : str =jax.nn.log_softmax(lowerCamelCase__ , axis=-1 ) lowerCamelCase_ : Tuple =-jnp.sum(labels * logits , axis=-1 ) if reduction is not None: lowerCamelCase_ : str =reduction(lowerCamelCase__ ) return loss lowerCamelCase_ : int =partial(lowerCamelCase__ , reduction=jnp.mean ) lowerCamelCase_ : int =cross_entropy(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ : Any =cross_entropy(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ : List[str] =cross_entropy(lowerCamelCase__ , lowerCamelCase__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class lowercase__ : _UpperCAmelCase :str = "google/bigbird-roberta-base" _UpperCAmelCase :int = 3000 _UpperCAmelCase :int = 10500 _UpperCAmelCase :int = 128 _UpperCAmelCase :int = 3 _UpperCAmelCase :int = 1 _UpperCAmelCase :int = 5 # tx_args _UpperCAmelCase :float = 3e-5 _UpperCAmelCase :float = 0.0 _UpperCAmelCase :int = 20000 _UpperCAmelCase :float = 0.00_95 _UpperCAmelCase :str = "bigbird-roberta-natural-questions" _UpperCAmelCase :str = "training-expt" _UpperCAmelCase :str = "data/nq-training.jsonl" _UpperCAmelCase :str = "data/nq-validation.jsonl" def UpperCAmelCase__ ( self : Union[str, Any] ): os.makedirs(self.base_dir , exist_ok=snake_case__ ) lowerCamelCase_ : Tuple =os.path.join(self.base_dir , self.save_dir ) lowerCamelCase_ : Optional[Any] =self.batch_size_per_device * jax.device_count() @dataclass class lowercase__ : _UpperCAmelCase :int _UpperCAmelCase :int = 4096 # no dynamic padding on TPUs def __call__( self : List[str] , snake_case__ : List[str] ): lowerCamelCase_ : Optional[int] =self.collate_fn(snake_case__ ) lowerCamelCase_ : List[str] =jax.tree_util.tree_map(snake_case__ , snake_case__ ) return batch def UpperCAmelCase__ ( self : str , snake_case__ : Dict ): lowerCamelCase_ , lowerCamelCase_ : Union[str, Any] =self.fetch_inputs(features["input_ids"] ) lowerCamelCase_ : Dict ={ "input_ids": jnp.array(snake_case__ , dtype=jnp.intaa ), "attention_mask": jnp.array(snake_case__ , dtype=jnp.intaa ), "start_labels": jnp.array(features["start_token"] , dtype=jnp.intaa ), "end_labels": jnp.array(features["end_token"] , dtype=jnp.intaa ), "pooled_labels": jnp.array(features["category"] , dtype=jnp.intaa ), } return batch def UpperCAmelCase__ ( self : List[Any] , snake_case__ : list ): lowerCamelCase_ : Any =[self._fetch_inputs(snake_case__ ) for ids in input_ids] return zip(*snake_case__ ) def UpperCAmelCase__ ( self : int , snake_case__ : list ): lowerCamelCase_ : List[Any] =[1 for _ in range(len(snake_case__ ) )] while len(snake_case__ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def _snake_case ( lowerCamelCase__ : int , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[int]=None ) -> Optional[int]: if seed is not None: lowerCamelCase_ : Union[str, Any] =dataset.shuffle(seed=lowerCamelCase__ ) for i in range(len(lowerCamelCase__ ) // batch_size ): lowerCamelCase_ : Any =dataset[i * batch_size : (i + 1) * batch_size] yield dict(lowerCamelCase__ ) @partial(jax.pmap , axis_name="batch" ) def _snake_case ( lowerCamelCase__ : Dict , lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Tuple ) -> int: def loss_fn(lowerCamelCase__ : Optional[int] ): lowerCamelCase_ : List[Any] =model_inputs.pop("start_labels" ) lowerCamelCase_ : Dict =model_inputs.pop("end_labels" ) lowerCamelCase_ : Any =model_inputs.pop("pooled_labels" ) lowerCamelCase_ : Tuple =state.apply_fn(**lowerCamelCase__ , params=lowerCamelCase__ , dropout_rng=lowerCamelCase__ , train=lowerCamelCase__ ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Any =outputs return state.loss_fn( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) lowerCamelCase_ , lowerCamelCase_ : Optional[Any] =jax.random.split(lowerCamelCase__ ) lowerCamelCase_ : Union[str, Any] =jax.value_and_grad(lowerCamelCase__ ) lowerCamelCase_ , lowerCamelCase_ : Tuple =grad_fn(state.params ) lowerCamelCase_ : List[Any] =jax.lax.pmean({"loss": loss} , axis_name="batch" ) lowerCamelCase_ : int =jax.lax.pmean(lowerCamelCase__ , "batch" ) lowerCamelCase_ : List[Any] =state.apply_gradients(grads=lowerCamelCase__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="batch" ) def _snake_case ( lowerCamelCase__ : List[str] , **lowerCamelCase__ : Union[str, Any] ) -> Dict: lowerCamelCase_ : Dict =model_inputs.pop("start_labels" ) lowerCamelCase_ : List[Any] =model_inputs.pop("end_labels" ) lowerCamelCase_ : Union[str, Any] =model_inputs.pop("pooled_labels" ) lowerCamelCase_ : Tuple =state.apply_fn(**lowerCamelCase__ , params=state.params , train=lowerCamelCase__ ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Union[str, Any] =outputs lowerCamelCase_ : int =state.loss_fn(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ : str =jax.lax.pmean({"loss": loss} , axis_name="batch" ) return metrics class lowercase__ ( train_state.TrainState ): _UpperCAmelCase :Callable = struct.field(pytree_node=snake_case__ ) @dataclass class lowercase__ : _UpperCAmelCase :Args _UpperCAmelCase :Callable _UpperCAmelCase :Callable _UpperCAmelCase :Callable _UpperCAmelCase :Callable _UpperCAmelCase :wandb _UpperCAmelCase :Callable = None def UpperCAmelCase__ ( self : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : Any , snake_case__ : List[str] , snake_case__ : str=None ): lowerCamelCase_ : int =model.params lowerCamelCase_ : Optional[Any] =TrainState.create( apply_fn=model.__call__ , params=snake_case__ , tx=snake_case__ , loss_fn=snake_case__ , ) if ckpt_dir is not None: lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Any =restore_checkpoint(snake_case__ , snake_case__ ) lowerCamelCase_ : Tuple ={ "lr": args.lr, "init_lr": args.init_lr, "warmup_steps": args.warmup_steps, "num_train_steps": num_train_steps, "weight_decay": args.weight_decay, } lowerCamelCase_ , lowerCamelCase_ : Tuple =build_tx(**snake_case__ ) lowerCamelCase_ : Union[str, Any] =train_state.TrainState( step=snake_case__ , apply_fn=model.__call__ , params=snake_case__ , tx=snake_case__ , opt_state=snake_case__ , ) lowerCamelCase_ : int =args lowerCamelCase_ : Union[str, Any] =data_collator lowerCamelCase_ : Dict =lr lowerCamelCase_ : Optional[Any] =params lowerCamelCase_ : Dict =jax_utils.replicate(snake_case__ ) return state def UpperCAmelCase__ ( self : Dict , snake_case__ : List[Any] , snake_case__ : int , snake_case__ : List[str] ): lowerCamelCase_ : str =self.args lowerCamelCase_ : List[Any] =len(snake_case__ ) // args.batch_size lowerCamelCase_ : Optional[int] =jax.random.PRNGKey(0 ) lowerCamelCase_ : Dict =jax.random.split(snake_case__ , jax.device_count() ) for epoch in range(args.max_epochs ): lowerCamelCase_ : int =jnp.array(0 , dtype=jnp.floataa ) lowerCamelCase_ : List[Any] =get_batched_dataset(snake_case__ , args.batch_size , seed=snake_case__ ) lowerCamelCase_ : Dict =0 for batch in tqdm(snake_case__ , total=snake_case__ , desc=F"""Running EPOCH-{epoch}""" ): lowerCamelCase_ : str =self.data_collator(snake_case__ ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Any =self.train_step_fn(snake_case__ , snake_case__ , **snake_case__ ) running_loss += jax_utils.unreplicate(metrics["loss"] ) i += 1 if i % args.logging_steps == 0: lowerCamelCase_ : Tuple =jax_utils.unreplicate(state.step ) lowerCamelCase_ : Optional[Any] =running_loss.item() / i lowerCamelCase_ : Any =self.scheduler_fn(state_step - 1 ) lowerCamelCase_ : Optional[Any] =self.evaluate(snake_case__ , snake_case__ ) lowerCamelCase_ : str ={ "step": state_step.item(), "eval_loss": eval_loss.item(), "tr_loss": tr_loss, "lr": lr.item(), } tqdm.write(str(snake_case__ ) ) self.logger.log(snake_case__ , commit=snake_case__ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + F"""-e{epoch}-s{i}""" , state=snake_case__ ) def UpperCAmelCase__ ( self : str , snake_case__ : Dict , snake_case__ : Union[str, Any] ): lowerCamelCase_ : List[Any] =get_batched_dataset(snake_case__ , self.args.batch_size ) lowerCamelCase_ : List[str] =len(snake_case__ ) // self.args.batch_size lowerCamelCase_ : Tuple =jnp.array(0 , dtype=jnp.floataa ) lowerCamelCase_ : Any =0 for batch in tqdm(snake_case__ , total=snake_case__ , desc="Evaluating ... " ): lowerCamelCase_ : Optional[Any] =self.data_collator(snake_case__ ) lowerCamelCase_ : List[str] =self.val_step_fn(snake_case__ , **snake_case__ ) running_loss += jax_utils.unreplicate(metrics["loss"] ) i += 1 return running_loss / i def UpperCAmelCase__ ( self : str , snake_case__ : Optional[int] , snake_case__ : Any ): lowerCamelCase_ : List[Any] =jax_utils.unreplicate(snake_case__ ) print(F"""SAVING CHECKPOINT IN {save_dir}""" , end=" ... " ) self.model_save_fn(snake_case__ , params=state.params ) with open(os.path.join(snake_case__ , "opt_state.msgpack" ) , "wb" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(snake_case__ , "args.joblib" ) ) joblib.dump(self.data_collator , os.path.join(snake_case__ , "data_collator.joblib" ) ) with open(os.path.join(snake_case__ , "training_state.json" ) , "w" ) as f: json.dump({"step": state.step.item()} , snake_case__ ) print("DONE" ) def _snake_case ( lowerCamelCase__ : List[str] , lowerCamelCase__ : Any ) -> List[Any]: print(F"""RESTORING CHECKPOINT FROM {save_dir}""" , end=" ... " ) with open(os.path.join(lowerCamelCase__ , "flax_model.msgpack" ) , "rb" ) as f: lowerCamelCase_ : Any =from_bytes(state.params , f.read() ) with open(os.path.join(lowerCamelCase__ , "opt_state.msgpack" ) , "rb" ) as f: lowerCamelCase_ : Optional[Any] =from_bytes(state.opt_state , f.read() ) lowerCamelCase_ : List[Any] =joblib.load(os.path.join(lowerCamelCase__ , "args.joblib" ) ) lowerCamelCase_ : int =joblib.load(os.path.join(lowerCamelCase__ , "data_collator.joblib" ) ) with open(os.path.join(lowerCamelCase__ , "training_state.json" ) , "r" ) as f: lowerCamelCase_ : Optional[Any] =json.load(lowerCamelCase__ ) lowerCamelCase_ : Optional[Any] =training_state["step"] print("DONE" ) return params, opt_state, step, args, data_collator def _snake_case ( lowerCamelCase__ : List[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Any ) -> str: lowerCamelCase_ : Dict =num_train_steps - warmup_steps lowerCamelCase_ : Optional[Any] =optax.linear_schedule(init_value=lowerCamelCase__ , end_value=lowerCamelCase__ , transition_steps=lowerCamelCase__ ) lowerCamelCase_ : List[Any] =optax.linear_schedule(init_value=lowerCamelCase__ , end_value=1e-7 , transition_steps=lowerCamelCase__ ) lowerCamelCase_ : Dict =optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def _snake_case ( lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Optional[int] ) -> List[str]: def weight_decay_mask(lowerCamelCase__ : str ): lowerCamelCase_ : Union[str, Any] =traverse_util.flatten_dict(lowerCamelCase__ ) lowerCamelCase_ : Any ={k: (v[-1] != "bias" and v[-2:] != ("LayerNorm", "scale")) for k, v in params.items()} return traverse_util.unflatten_dict(lowerCamelCase__ ) lowerCamelCase_ : Dict =scheduler_fn(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ : List[str] =optax.adamw(learning_rate=lowerCamelCase__ , weight_decay=lowerCamelCase__ , mask=lowerCamelCase__ ) return tx, lr
144
0
import os from datetime import datetime as dt from github import Github _snake_case = [ "good first issue", "feature request", "wip", ] def A ( ): '''simple docstring''' _lowerCAmelCase : Tuple = Github(os.environ["GITHUB_TOKEN"] ) _lowerCAmelCase : Tuple = g.get_repo("huggingface/accelerate" ) _lowerCAmelCase : str = repo.get_issues(state="open" ) for issue in open_issues: _lowerCAmelCase : str = sorted([comment for comment in issue.get_comments()] , key=lambda _lowerCamelCase : i.created_at , reverse=_lowerCamelCase ) _lowerCAmelCase : Union[str, Any] = comments[0] if len(_lowerCamelCase ) > 0 else None _lowerCAmelCase : Union[str, Any] = dt.utcnow() _lowerCAmelCase : str = (current_time - issue.updated_at).days _lowerCAmelCase : List[str] = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state="closed" ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
300
from __future__ import annotations from typing import Any class UpperCAmelCase_ : def __init__( self, __a, __a, __a = 0): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : int = row, column _lowerCAmelCase : str = [[default_value for c in range(__a)] for r in range(__a)] def __str__( self): '''simple docstring''' _lowerCAmelCase : Tuple = f"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier _lowerCAmelCase : str = 0 for row_vector in self.array: for obj in row_vector: _lowerCAmelCase : List[str] = max(__a, len(str(__a))) _lowerCAmelCase : Union[str, Any] = f"%{max_element_length}s" # Make string and return def single_line(__a) -> str: nonlocal string_format_identifier _lowerCAmelCase : Dict = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector) line += "]" return line s += "\n".join(single_line(__a) for row_vector in self.array) return s def __repr__( self): '''simple docstring''' return str(self) def snake_case__ ( self, __a): '''simple docstring''' if not (isinstance(__a, (list, tuple)) and len(__a) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, __a): '''simple docstring''' assert self.validate_indicies(__a) return self.array[loc[0]][loc[1]] def __setitem__( self, __a, __a): '''simple docstring''' assert self.validate_indicies(__a) _lowerCAmelCase : Union[str, Any] = value def __add__( self, __a): '''simple docstring''' assert isinstance(__a, __a) assert self.row == another.row and self.column == another.column # Add _lowerCAmelCase : Any = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowerCAmelCase : Any = self[r, c] + another[r, c] return result def __neg__( self): '''simple docstring''' _lowerCAmelCase : List[str] = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowerCAmelCase : str = -self[r, c] return result def __sub__( self, __a): '''simple docstring''' return self + (-another) def __mul__( self, __a): '''simple docstring''' if isinstance(__a, (int, float)): # Scalar multiplication _lowerCAmelCase : Dict = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowerCAmelCase : Optional[Any] = self[r, c] * another return result elif isinstance(__a, __a): # Matrix multiplication assert self.column == another.row _lowerCAmelCase : List[str] = Matrix(self.row, another.column) for r in range(self.row): for c in range(another.column): for i in range(self.column): result[r, c] += self[r, i] * another[i, c] return result else: _lowerCAmelCase : Optional[Any] = f"Unsupported type given for another ({type(__a)})" raise TypeError(__a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[Any] = Matrix(self.column, self.row) for r in range(self.row): for c in range(self.column): _lowerCAmelCase : Any = self[r, c] return result def snake_case__ ( self, __a, __a): '''simple docstring''' assert isinstance(__a, __a) and isinstance(__a, __a) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowerCAmelCase : int = v.transpose() _lowerCAmelCase : str = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def A ( ): '''simple docstring''' _lowerCAmelCase : List[Any] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowerCAmelCase : Union[str, Any] = 1 print(F"a^(-1) is {ainv}" ) # u, v _lowerCAmelCase : Any = Matrix(3 , 1 , 0 ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = 1, 2, -3 _lowerCAmelCase : List[Any] = Matrix(3 , 1 , 0 ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : str = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(_lowerCamelCase , _lowerCamelCase )}" ) def A ( ): '''simple docstring''' import doctest doctest.testmod() testa()
300
1
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor UpperCAmelCase__ = logging.get_logger(__name__) class lowercase_ ( lowercase ): '''simple docstring''' def __init__( self : Union[str, Any] , *__UpperCAmelCase : str , **__UpperCAmelCase : Tuple ) ->None: """simple docstring""" warnings.warn( '''The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use OwlViTImageProcessor instead.''' , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
0
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) _snake_case = parser.parse_args() _snake_case = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) _snake_case = CLIPImageProcessor() _snake_case = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') _snake_case = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
157
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.bpe.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model''', } } lowerCAmelCase__ = { '''camembert-base''': 512, } lowerCAmelCase__ = '''▁''' class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Dict = ['input_ids', 'attention_mask'] def __init__( self : Dict ,lowercase__ : Dict ,lowercase__ : Union[str, Any]="<s>" ,lowercase__ : List[str]="</s>" ,lowercase__ : Optional[int]="</s>" ,lowercase__ : List[Any]="<s>" ,lowercase__ : str="<unk>" ,lowercase__ : Dict="<pad>" ,lowercase__ : Any="<mask>" ,lowercase__ : Union[str, Any]=["<s>NOTUSED", "</s>NOTUSED"] ,lowercase__ : Optional[Dict[str, Any]] = None ,**lowercase__ : Dict ,): # Mask token behave like a normal word, i.e. include the space before it __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else mask_token __lowercase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,cls_token=lowercase__ ,pad_token=lowercase__ ,mask_token=lowercase__ ,additional_special_tokens=lowercase__ ,sp_model_kwargs=self.sp_model_kwargs ,**lowercase__ ,) __lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowercase__ ) ) __lowercase = vocab_file # HACK: These tokens were added by fairseq but don't seem to be actually used when duplicated in the actual # sentencepiece vocabulary (this is the case for <s> and </s> __lowercase = {'''<s>NOTUSED''': 0, '''<pad>''': 1, '''</s>NOTUSED''': 2, '''<unk>''': 3} __lowercase = len(self.fairseq_tokens_to_ids ) __lowercase = len(self.sp_model ) + len(self.fairseq_tokens_to_ids ) __lowercase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowercase = [self.cls_token_id] __lowercase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase__ ,token_ids_a=lowercase__ ,already_has_special_tokens=lowercase__ ) if token_ids_a is None: return [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1, 1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return len(self.fairseq_tokens_to_ids ) + len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : str ): return self.sp_model.encode(lowercase__ ,out_type=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Dict ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] elif self.sp_model.PieceToId(lowercase__ ) == 0: # Convert sentence piece unk token to fairseq unk token index return self.unk_token_id return self.fairseq_offset + self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : int ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : int ): __lowercase = [] __lowercase = '''''' __lowercase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase__ ) + token __lowercase = True __lowercase = [] else: current_sub_tokens.append(lowercase__ ) __lowercase = False out_string += self.sp_model.decode(lowercase__ ) return out_string.strip() def __getstate__( self : Union[str, Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : Union[str, Any] ,lowercase__ : List[str] ): __lowercase = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs''' ): __lowercase = {} __lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,lowercase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowercase__ ,'''wb''' ) as fi: __lowercase = self.sp_model.serialized_model_proto() fi.write(lowercase__ ) return (out_vocab_file,)
52
'''simple docstring''' def _A ( A__ ): """simple docstring""" stooge(A__ , 0 , len(A__ ) - 1 ) return arr def _A ( A__ , A__ , A__ ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: __lowercase , __lowercase = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: __lowercase = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(A__ , A__ , (h - t) ) # Recursively sort last 2/3 elements stooge(A__ , i + t , (A__) ) # Recursively sort first 2/3 elements stooge(A__ , A__ , (h - t) ) if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
52
1
import heapq as hq import math from collections.abc import Iterator class lowercase : def __init__( self ,A__): lowercase = str(id_) lowercase = None lowercase = None lowercase = [] lowercase = {} # {vertex:distance} def __lt__( self ,A__): return self.key < other.key def __repr__( self): return self.id def A__ ( self ,A__): self.neighbors.append(A__) def A__ ( self ,A__ ,A__): lowercase = weight def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , lowerCAmelCase__ ) graph[b - 1].add_edge(graph[a - 1] , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for u in graph: lowercase = math.inf lowercase = None lowercase = 0 lowercase = graph[:] while q: lowercase = min(lowerCAmelCase__ ) q.remove(lowerCAmelCase__ ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): lowercase = u lowercase = u.edges[v.id] for i in range(1 , len(lowerCAmelCase__ ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for u in graph: lowercase = math.inf lowercase = None lowercase = 0 lowercase = list(lowerCAmelCase__ ) hq.heapify(lowerCAmelCase__ ) while h: lowercase = hq.heappop(lowerCAmelCase__ ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): lowercase = u lowercase = u.edges[v.id] hq.heapify(lowerCAmelCase__ ) for i in range(1 , len(lowerCAmelCase__ ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCamelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
101
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : List[Any] = logging.get_logger(__name__) _lowerCAmelCase : List[str] = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class _UpperCamelCase ( lowerCAmelCase ): UpperCAmelCase_ = """realm""" def __init__( self :str , lowerCamelCase :List[Any]=3_0522 , lowerCamelCase :Optional[int]=768 , lowerCamelCase :Any=128 , lowerCamelCase :Tuple=12 , lowerCamelCase :str=12 , lowerCamelCase :List[str]=8 , lowerCamelCase :List[str]=3072 , lowerCamelCase :List[str]="gelu_new" , lowerCamelCase :int=0.1 , lowerCamelCase :Optional[Any]=0.1 , lowerCamelCase :int=512 , lowerCamelCase :Union[str, Any]=2 , lowerCamelCase :str=0.02 , lowerCamelCase :Tuple=1e-12 , lowerCamelCase :Dict=256 , lowerCamelCase :int=10 , lowerCamelCase :List[str]=1e-3 , lowerCamelCase :str=5 , lowerCamelCase :Optional[int]=320 , lowerCamelCase :Union[str, Any]=1335_3718 , lowerCamelCase :str=5000 , lowerCamelCase :str=1 , lowerCamelCase :List[Any]=0 , lowerCamelCase :Tuple=2 , **lowerCamelCase :Optional[int] , ) -> Optional[Any]: super().__init__(pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , **lowerCamelCase ) # Common config UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = hidden_size UpperCAmelCase__ = retriever_proj_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = num_candidates UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = layer_norm_eps # Reader config UpperCAmelCase__ = span_hidden_size UpperCAmelCase__ = max_span_width UpperCAmelCase__ = reader_layer_norm_eps UpperCAmelCase__ = reader_beam_size UpperCAmelCase__ = reader_seq_len # Retrieval config UpperCAmelCase__ = num_block_records UpperCAmelCase__ = searcher_beam_size
169
0
import os import string import sys lowercase_ = 1 << 8 lowercase_ = { 'tab': ord('\t'), 'newline': ord('\r'), 'esc': 2_7, 'up': 6_5 + ARROW_KEY_FLAG, 'down': 6_6 + ARROW_KEY_FLAG, 'right': 6_7 + ARROW_KEY_FLAG, 'left': 6_8 + ARROW_KEY_FLAG, 'mod_int': 9_1, 'undefined': sys.maxsize, 'interrupt': 3, 'insert': 5_0, 'delete': 5_1, 'pg_up': 5_3, 'pg_down': 5_4, } lowercase_ = KEYMAP['up'] lowercase_ = KEYMAP['left'] if sys.platform == "win32": lowercase_ = [] lowercase_ = { b'\xe0H': KEYMAP['up'] - ARROW_KEY_FLAG, b'\x00H': KEYMAP['up'] - ARROW_KEY_FLAG, b'\xe0P': KEYMAP['down'] - ARROW_KEY_FLAG, b'\x00P': KEYMAP['down'] - ARROW_KEY_FLAG, b'\xe0M': KEYMAP['right'] - ARROW_KEY_FLAG, b'\x00M': KEYMAP['right'] - ARROW_KEY_FLAG, b'\xe0K': KEYMAP['left'] - ARROW_KEY_FLAG, b'\x00K': KEYMAP['left'] - ARROW_KEY_FLAG, } for i in range(1_0): lowercase_ = ord(str(i)) def UpperCamelCase__ ( ): if os.name == "nt": import msvcrt __lowerCamelCase : Dict = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(SCREAMING_SNAKE_CASE__ ) == 0: # Read the keystroke __lowerCamelCase : Tuple = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): __lowerCamelCase : Tuple = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: __lowerCamelCase : Tuple = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['mod_int'] ) ) WIN_CH_BUFFER.append(SCREAMING_SNAKE_CASE__ ) if ord(SCREAMING_SNAKE_CASE__ ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) __lowerCamelCase : Optional[int] = chr(KEYMAP['esc'] ) except KeyError: __lowerCamelCase : Tuple = cha[1] else: __lowerCamelCase : str = ch.decode(SCREAMING_SNAKE_CASE__ ) else: __lowerCamelCase : Optional[Any] = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty __lowerCamelCase : List[str] = sys.stdin.fileno() __lowerCamelCase : Dict = termios.tcgetattr(SCREAMING_SNAKE_CASE__ ) try: tty.setraw(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Dict = sys.stdin.read(1 ) finally: termios.tcsetattr(SCREAMING_SNAKE_CASE__ , termios.TCSADRAIN , SCREAMING_SNAKE_CASE__ ) return ch def UpperCamelCase__ ( ): __lowerCamelCase : Any = get_raw_chars() if ord(SCREAMING_SNAKE_CASE__ ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(SCREAMING_SNAKE_CASE__ ) == KEYMAP["esc"]: __lowerCamelCase : str = get_raw_chars() if ord(SCREAMING_SNAKE_CASE__ ) == KEYMAP["mod_int"]: __lowerCamelCase : Optional[int] = get_raw_chars() if ord(SCREAMING_SNAKE_CASE__ ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(SCREAMING_SNAKE_CASE__ ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(SCREAMING_SNAKE_CASE__ ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
350
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration lowercase_ = 5_0_0_0_0_0 lowercase_ ,lowercase_ = os.path.split(__file__) lowercase_ = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Optional[int] = dataset.map(**SCREAMING_SNAKE_CASE__ ) @get_duration def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Tuple = dataset.filter(**SCREAMING_SNAKE_CASE__ ) def UpperCamelCase__ ( ): __lowerCamelCase : str = {'num examples': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Any = datasets.Features({'text': datasets.Value('string' ), 'numbers': datasets.Value('float32' )} ) __lowerCamelCase : Any = generate_example_dataset( os.path.join(SCREAMING_SNAKE_CASE__ , 'dataset.arrow' ) , SCREAMING_SNAKE_CASE__ , num_examples=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : List[Any] = transformers.AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=SCREAMING_SNAKE_CASE__ ) def tokenize(SCREAMING_SNAKE_CASE__ ): return tokenizer(examples['text'] ) __lowerCamelCase : str = map(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : int = map(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[int] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='numpy' ): __lowerCamelCase : Union[str, Any] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='pandas' ): __lowerCamelCase : Any = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='torch' , columns='numbers' ): __lowerCamelCase : List[str] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='tensorflow' , columns='numbers' ): __lowerCamelCase : List[Any] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[Any] = map(SCREAMING_SNAKE_CASE__ , function=SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[int] = filter(SCREAMING_SNAKE_CASE__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(SCREAMING_SNAKE_CASE__ , 'wb' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ).encode('utf-8' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
194
0
'''simple docstring''' def snake_case_ (_a : int , _a : int ): while b: UpperCAmelCase , UpperCAmelCase = b, a % b return a def snake_case_ (_a : int , _a : int ): return a if b == 0 else euclidean_gcd_recursive(_a , a % b ) def snake_case_ (): print(F"euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}" ) print(F"euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}" ) print(F"euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}" ) print(F"euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}" ) print(F"euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}" ) print(F"euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}" ) print(F"euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}" ) print(F"euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}" ) print(F"euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}" ) print(F"euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}" ) if __name__ == "__main__": main()
34
'''simple docstring''' from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class _a ( __a ): __a : str = ["""vqvae"""] def __init__( self : str , lowercase : AutoencoderKL , lowercase : UNetaDConditionModel , lowercase : Mel , lowercase : Union[DDIMScheduler, DDPMScheduler] , ): '''simple docstring''' super().__init__() self.register_modules(unet=lowercase , scheduler=lowercase , mel=lowercase , vqvae=lowercase ) def A ( self : Optional[Any] ): '''simple docstring''' return 50 if isinstance(self.scheduler , lowercase ) else 1_000 @torch.no_grad() def __call__( self : Optional[Any] , lowercase : int = 1 , lowercase : str = None , lowercase : np.ndarray = None , lowercase : int = 0 , lowercase : int = 0 , lowercase : int = None , lowercase : torch.Generator = None , lowercase : float = 0 , lowercase : float = 0 , lowercase : torch.Generator = None , lowercase : float = 0 , lowercase : torch.Tensor = None , lowercase : torch.Tensor = None , lowercase : Tuple=True , ): '''simple docstring''' UpperCAmelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(lowercase ) UpperCAmelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: UpperCAmelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: UpperCAmelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=lowercase , device=self.device , ) UpperCAmelCase = noise UpperCAmelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(lowercase , lowercase ) UpperCAmelCase = self.mel.audio_slice_to_image(lowercase ) UpperCAmelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) UpperCAmelCase = (input_image / 255) * 2 - 1 UpperCAmelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: UpperCAmelCase = self.vqvae.encode(torch.unsqueeze(lowercase , 0 ) ).latent_dist.sample( generator=lowercase )[0] UpperCAmelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: UpperCAmelCase = self.scheduler.add_noise(lowercase , lowercase , self.scheduler.timesteps[start_step - 1] ) UpperCAmelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) UpperCAmelCase = int(mask_start_secs * pixels_per_second ) UpperCAmelCase = int(mask_end_secs * pixels_per_second ) UpperCAmelCase = self.scheduler.add_noise(lowercase , lowercase , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , lowercase ): UpperCAmelCase = self.unet(lowercase , lowercase , lowercase )['''sample'''] else: UpperCAmelCase = self.unet(lowercase , lowercase )['''sample'''] if isinstance(self.scheduler , lowercase ): UpperCAmelCase = self.scheduler.step( model_output=lowercase , timestep=lowercase , sample=lowercase , eta=lowercase , generator=lowercase , )['''prev_sample'''] else: UpperCAmelCase = self.scheduler.step( model_output=lowercase , timestep=lowercase , sample=lowercase , generator=lowercase , )['''prev_sample'''] if mask is not None: if mask_start > 0: UpperCAmelCase = mask[:, step, :, :mask_start] if mask_end > 0: UpperCAmelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance UpperCAmelCase = 1 / self.vqvae.config.scaling_factor * images UpperCAmelCase = self.vqvae.decode(lowercase )['''sample'''] UpperCAmelCase = (images / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() UpperCAmelCase = (images * 255).round().astype('''uint8''' ) UpperCAmelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(lowercase , mode='''RGB''' ).convert('''L''' ) for _ in images) ) UpperCAmelCase = [self.mel.image_to_audio(lowercase ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(lowercase )[:, np.newaxis, :] ) , **ImagePipelineOutput(lowercase ) ) @torch.no_grad() def A ( self : Dict , lowercase : List[Image.Image] , lowercase : int = 50 ): '''simple docstring''' assert isinstance(self.scheduler , lowercase ) self.scheduler.set_timesteps(lowercase ) UpperCAmelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) UpperCAmelCase = (sample / 255) * 2 - 1 UpperCAmelCase = torch.Tensor(lowercase ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): UpperCAmelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps UpperCAmelCase = self.scheduler.alphas_cumprod[t] UpperCAmelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) UpperCAmelCase = 1 - alpha_prod_t UpperCAmelCase = self.unet(lowercase , lowercase )['''sample'''] UpperCAmelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output UpperCAmelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) UpperCAmelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def A ( lowercase : torch.Tensor , lowercase : torch.Tensor , lowercase : float ): '''simple docstring''' UpperCAmelCase = acos(torch.dot(torch.flatten(lowercase ) , torch.flatten(lowercase ) ) / torch.norm(lowercase ) / torch.norm(lowercase ) ) return sin((1 - alpha) * theta ) * xa / sin(lowercase ) + sin(alpha * theta ) * xa / sin(lowercase )
34
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : List[Any] = logging.get_logger(__name__) class __A ( lowerCAmelCase ): lowerCAmelCase_ : List[str] = "timm_backbone" def __init__( self : Dict , UpperCAmelCase_ : List[Any]=None , UpperCAmelCase_ : List[str]=3 , UpperCAmelCase_ : str=True , UpperCAmelCase_ : str=True , UpperCAmelCase_ : Tuple=None , **UpperCAmelCase_ : List[Any] , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase : Tuple = backbone lowerCAmelCase : Optional[Any] = num_channels lowerCAmelCase : List[Any] = features_only lowerCAmelCase : Union[str, Any] = use_pretrained_backbone lowerCAmelCase : List[str] = True lowerCAmelCase : str = out_indices if out_indices is not None else (-1,)
323
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> List[Any]: '''simple docstring''' return x + 2 class __A ( unittest.TestCase ): def lowercase__ ( self : int ): lowerCAmelCase : List[str] = 'x = 3' lowerCAmelCase : Optional[Any] = {} lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) assert result == 3 self.assertDictEqual(UpperCAmelCase_ , {'x': 3} ) lowerCAmelCase : Dict = 'x = y' lowerCAmelCase : List[Any] = {'y': 5} lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 5, 'y': 5} ) def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : Any = 'y = add_two(x)' lowerCAmelCase : int = {'x': 3} lowerCAmelCase : Optional[int] = evaluate(UpperCAmelCase_ , {'add_two': add_two} , state=UpperCAmelCase_ ) assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'y': 5} ) # Won't work without the tool with CaptureStdout() as out: lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) assert result is None assert "tried to execute add_two" in out.out def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Tuple = 'x = 3' lowerCAmelCase : List[Any] = {} lowerCAmelCase : Dict = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) assert result == 3 self.assertDictEqual(UpperCAmelCase_ , {'x': 3} ) def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : List[Any] = 'test_dict = {\'x\': x, \'y\': add_two(x)}' lowerCAmelCase : Dict = {'x': 3} lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {'add_two': add_two} , state=UpperCAmelCase_ ) self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'y': 5} ) self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'test_dict': {'x': 3, 'y': 5}} ) def lowercase__ ( self : Any ): lowerCAmelCase : Union[str, Any] = 'x = 3\ny = 5' lowerCAmelCase : str = {} lowerCAmelCase : Optional[int] = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'y': 5} ) def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Union[str, Any] = 'text = f\'This is x: {x}.\'' lowerCAmelCase : str = {'x': 3} lowerCAmelCase : int = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'text': 'This is x: 3.'} ) def lowercase__ ( self : Dict ): lowerCAmelCase : Optional[Any] = 'if x <= 3:\n y = 2\nelse:\n y = 5' lowerCAmelCase : Dict = {'x': 3} lowerCAmelCase : int = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'y': 2} ) lowerCAmelCase : Any = {'x': 8} lowerCAmelCase : Optional[int] = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 8, 'y': 5} ) def lowercase__ ( self : List[Any] ): lowerCAmelCase : int = 'test_list = [x, add_two(x)]' lowerCAmelCase : Optional[Any] = {'x': 3} lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {'add_two': add_two} , state=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , [3, 5] ) self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'test_list': [3, 5]} ) def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : int = 'y = x' lowerCAmelCase : Optional[int] = {'x': 3} lowerCAmelCase : Tuple = evaluate(UpperCAmelCase_ , {} , state=UpperCAmelCase_ ) assert result == 3 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'y': 3} ) def lowercase__ ( self : List[str] ): lowerCAmelCase : Dict = 'test_list = [x, add_two(x)]\ntest_list[1]' lowerCAmelCase : List[str] = {'x': 3} lowerCAmelCase : List[str] = evaluate(UpperCAmelCase_ , {'add_two': add_two} , state=UpperCAmelCase_ ) assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'test_list': [3, 5]} ) lowerCAmelCase : Optional[Any] = 'test_dict = {\'x\': x, \'y\': add_two(x)}\ntest_dict[\'y\']' lowerCAmelCase : List[Any] = {'x': 3} lowerCAmelCase : Optional[Any] = evaluate(UpperCAmelCase_ , {'add_two': add_two} , state=UpperCAmelCase_ ) assert result == 5 self.assertDictEqual(UpperCAmelCase_ , {'x': 3, 'test_dict': {'x': 3, 'y': 5}} ) def lowercase__ ( self : int ): lowerCAmelCase : Any = 'x = 0\nfor i in range(3):\n x = i' lowerCAmelCase : str = {} lowerCAmelCase : Dict = evaluate(UpperCAmelCase_ , {'range': range} , state=UpperCAmelCase_ ) assert result == 2 self.assertDictEqual(UpperCAmelCase_ , {'x': 2, 'i': 2} )
323
1