code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, is_batched, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = ["""pixel_values"""] def __init__( self , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = PILImageResampling.BICUBIC , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = 1 / 255 , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> None: super().__init__(**lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = size if size is not None else {'height': 224, 'width': 224} SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else {'height': 224, 'width': 224} SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__ , param_name='crop_size' ) SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = do_rescale SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = do_center_crop SCREAMING_SNAKE_CASE = crop_size SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = resample SCREAMING_SNAKE_CASE = rescale_factor SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN SCREAMING_SNAKE_CASE = image_std if image_std is not None else IMAGENET_DEFAULT_STD def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = PILImageResampling.BILINEAR , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ ) if "shortest_edge" in size: SCREAMING_SNAKE_CASE = get_resize_output_image_size(lowerCAmelCase__ , size=size['shortest_edge'] , default_to_square=lowerCAmelCase__ ) # size = get_resize_output_image_size(image, size["shortest_edge"], size["longest_edge"]) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE = (size['height'], size['width']) else: raise ValueError(F'Size must contain \'height\' and \'width\' keys or \'shortest_edge\' key. Got {size.keys()}' ) return resize(lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(lowerCAmelCase__ , size=(size['height'], size['width']) , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> np.ndarray: return rescale(lowerCAmelCase__ , scale=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: return normalize(lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = ChannelDimension.FIRST , **lowerCAmelCase__ , ) -> BatchFeature: SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ , param_name='crop_size' , default_to_square=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE = size if size is not None else self.size SCREAMING_SNAKE_CASE = get_size_dict(lowerCAmelCase__ ) if not is_batched(lowerCAmelCase__ ): SCREAMING_SNAKE_CASE = [images] if not valid_images(lowerCAmelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(lowerCAmelCase__ ) for image in images] if do_resize: SCREAMING_SNAKE_CASE = [self.resize(image=lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE = [self.center_crop(image=lowerCAmelCase__ , size=lowerCAmelCase__ ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE = [self.rescale(image=lowerCAmelCase__ , scale=lowerCAmelCase__ ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE = [self.normalize(image=lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ ) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__ ) for image in images] SCREAMING_SNAKE_CASE = {'pixel_values': images} return BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__ )
247
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class lowerCAmelCase ( ctypes.Structure ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Any = [("""size""", ctypes.c_int), ("""visible""", ctypes.c_byte)] def lowercase () -> Optional[int]: if os.name == "nt": SCREAMING_SNAKE_CASE = CursorInfo() SCREAMING_SNAKE_CASE = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = False ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def lowercase () -> int: if os.name == "nt": SCREAMING_SNAKE_CASE = CursorInfo() SCREAMING_SNAKE_CASE = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = True ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def lowercase () -> Dict: try: hide_cursor() yield finally: show_cursor()
247
1
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def __lowerCamelCase ( *__a :List[str] , __a :Optional[Union[Dict, Any]] = None , __a :List[str]=True , __a :str=2 ) -> Union[str, Any]: """simple docstring""" from .. import __version__ A__ = take_from A__ = () if not isinstance(args[0] , __a ): A__ = (args,) for attribute, version_name, message in args: if version.parse(version.parse(__a ).base_version ) >= version.parse(__a ): raise ValueError( F'The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers\'' F' version {__version__} is >= {version_name}' ) A__ = None if isinstance(__a , __a ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(__a ),) A__ = F'The `{attribute}` argument is deprecated and will be removed in version {version_name}.' elif hasattr(__a , __a ): values += (getattr(__a , __a ),) A__ = F'The `{attribute}` attribute is deprecated and will be removed in version {version_name}.' elif deprecated_kwargs is None: A__ = F'`{attribute}` is deprecated and will be removed in version {version_name}.' if warning is not None: A__ = warning + """ """ if standard_warn else """""" warnings.warn(warning + message , __a , stacklevel=__a ) if isinstance(__a , __a ) and len(__a ) > 0: A__ = inspect.getouterframes(inspect.currentframe() )[1] A__ = call_frame.filename A__ = call_frame.lineno A__ = call_frame.function A__ , A__ = next(iter(deprecated_kwargs.items() ) ) raise TypeError(F'{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`' ) if len(__a ) == 0: return elif len(__a ) == 1: return values[0] return values
702
def __lowerCamelCase ( __a :int ) -> list[int]: """simple docstring""" if num <= 0: raise ValueError("""Input must be a positive integer""" ) A__ = [True] * (num + 1) A__ = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __a ): A__ = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() A : Any = int(input('''Enter a positive integer: ''').strip()) print(prime_sieve_eratosthenes(user_num))
247
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer __UpperCamelCase : Optional[Any] = logging.get_logger(__name__) __UpperCamelCase : Union[str, Any] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} __UpperCamelCase : Union[str, Any] = { 'vocab_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/vocab.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/vocab.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/vocab.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json' ), }, 'merges_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/merges.txt', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/merges.txt', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/merges.txt', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt' ), }, 'tokenizer_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/tokenizer.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/tokenizer.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json', 'roberta-base-openai-detector': ( 'https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json' ), 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json' ), }, } __UpperCamelCase : Union[str, Any] = { 'roberta-base': 512, 'roberta-large': 512, 'roberta-large-mnli': 512, 'distilroberta-base': 512, 'roberta-base-openai-detector': 512, 'roberta-large-openai-detector': 512, } class _UpperCamelCase ( A ): '''simple docstring''' a_ : Dict = VOCAB_FILES_NAMES a_ : Tuple = PRETRAINED_VOCAB_FILES_MAP a_ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : List[Any] = ["input_ids", "attention_mask"] a_ : Dict = RobertaTokenizer def __init__( self : Any , _lowerCamelCase : List[str]=None , _lowerCamelCase : Tuple=None , _lowerCamelCase : Optional[int]=None , _lowerCamelCase : Tuple="replace" , _lowerCamelCase : Optional[int]="<s>" , _lowerCamelCase : Any="</s>" , _lowerCamelCase : Optional[int]="</s>" , _lowerCamelCase : str="<s>" , _lowerCamelCase : Dict="<unk>" , _lowerCamelCase : Optional[Any]="<pad>" , _lowerCamelCase : Any="<mask>" , _lowerCamelCase : Any=False , _lowerCamelCase : Any=True , **_lowerCamelCase : List[Any] , ): '''simple docstring''' super().__init__( _lowerCamelCase , _lowerCamelCase , tokenizer_file=_lowerCamelCase , errors=_lowerCamelCase , bos_token=_lowerCamelCase , eos_token=_lowerCamelCase , sep_token=_lowerCamelCase , cls_token=_lowerCamelCase , unk_token=_lowerCamelCase , pad_token=_lowerCamelCase , mask_token=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase , **_lowerCamelCase , ) __lowerCamelCase : str = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , _lowerCamelCase ) != add_prefix_space: __lowerCamelCase : Tuple = getattr(_lowerCamelCase , pre_tok_state.pop("""type""" ) ) __lowerCamelCase : Union[str, Any] = add_prefix_space __lowerCamelCase : List[str] = pre_tok_class(**_lowerCamelCase ) __lowerCamelCase : int = add_prefix_space __lowerCamelCase : List[Any] = """post_processor""" __lowerCamelCase : Any = getattr(self.backend_tokenizer , _lowerCamelCase , _lowerCamelCase ) if tokenizer_component_instance: __lowerCamelCase : Tuple = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __lowerCamelCase : str = tuple(state["""sep"""] ) if "cls" in state: __lowerCamelCase : Dict = tuple(state["""cls"""] ) __lowerCamelCase : Optional[Any] = False if state.get("""add_prefix_space""" , _lowerCamelCase ) != add_prefix_space: __lowerCamelCase : Optional[Any] = add_prefix_space __lowerCamelCase : Dict = True if state.get("""trim_offsets""" , _lowerCamelCase ) != trim_offsets: __lowerCamelCase : Union[str, Any] = trim_offsets __lowerCamelCase : Optional[int] = True if changes_to_apply: __lowerCamelCase : List[str] = getattr(_lowerCamelCase , state.pop("""type""" ) ) __lowerCamelCase : Tuple = component_class(**_lowerCamelCase ) setattr(self.backend_tokenizer , _lowerCamelCase , _lowerCamelCase ) @property def _snake_case ( self : Any ): '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""" ) return None return str(self._mask_token ) @mask_token.setter def _snake_case ( self : Optional[Any] , _lowerCamelCase : Optional[Any] ): '''simple docstring''' __lowerCamelCase : Any = AddedToken(_lowerCamelCase , lstrip=_lowerCamelCase , rstrip=_lowerCamelCase ) if isinstance(_lowerCamelCase , _lowerCamelCase ) else value __lowerCamelCase : Union[str, Any] = value def _snake_case ( self : str , *_lowerCamelCase : Dict , **_lowerCamelCase : int ): '''simple docstring''' __lowerCamelCase : str = kwargs.get("""is_split_into_words""" , _lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*_lowerCamelCase , **_lowerCamelCase ) def _snake_case ( self : Dict , *_lowerCamelCase : Any , **_lowerCamelCase : str ): '''simple docstring''' __lowerCamelCase : Tuple = kwargs.get("""is_split_into_words""" , _lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*_lowerCamelCase , **_lowerCamelCase ) def _snake_case ( self : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : Optional[str] = None ): '''simple docstring''' __lowerCamelCase : Tuple = self._tokenizer.model.save(_lowerCamelCase , name=_lowerCamelCase ) return tuple(_lowerCamelCase ) def _snake_case ( self : Optional[Any] , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : List[str]=None ): '''simple docstring''' __lowerCamelCase : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _snake_case ( self : List[Any] , _lowerCamelCase : List[int] , _lowerCamelCase : Optional[List[int]] = None ): '''simple docstring''' __lowerCamelCase : Optional[Any] = [self.sep_token_id] __lowerCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
519
__UpperCamelCase : Optional[int] = 'Input must be a string of 8 numbers plus letter' __UpperCamelCase : Optional[Any] = 'TRWAGMYFPDXBNJZSQVHLCKE' def _UpperCAmelCase ( UpperCAmelCase : str ): """simple docstring""" if not isinstance(UpperCAmelCase , UpperCAmelCase ): __lowerCamelCase : Union[str, Any] = f"""Expected string as input, found {type(UpperCAmelCase ).__name__}""" raise TypeError(UpperCAmelCase ) __lowerCamelCase : List[Any] = spanish_id.replace("""-""" , """""" ).upper() if len(UpperCAmelCase ) != 9: raise ValueError(UpperCAmelCase ) try: __lowerCamelCase : Tuple = int(spanish_id_clean[0:8] ) __lowerCamelCase : Optional[int] = spanish_id_clean[8] except ValueError as ex: raise ValueError(UpperCAmelCase ) from ex if letter.isdigit(): raise ValueError(UpperCAmelCase ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
519
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices a_ = logging.get_logger(__name__) a_ = { '''shi-labs/dinat-mini-in1k-224''': '''https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json''', # See all Dinat models at https://huggingface.co/models?filter=dinat } class __SCREAMING_SNAKE_CASE ( snake_case__ , snake_case__ ): snake_case_ = '''dinat''' snake_case_ = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : List[Any] , __lowercase : Optional[int]=4 , __lowercase : int=3 , __lowercase : Tuple=64 , __lowercase : int=[3, 4, 6, 5] , __lowercase : int=[2, 4, 8, 16] , __lowercase : str=7 , __lowercase : Dict=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] , __lowercase : List[Any]=3.0 , __lowercase : int=True , __lowercase : Optional[Any]=0.0 , __lowercase : Union[str, Any]=0.0 , __lowercase : str=0.1 , __lowercase : List[Any]="gelu" , __lowercase : str=0.02 , __lowercase : str=1e-5 , __lowercase : Any=0.0 , __lowercase : Dict=None , __lowercase : Any=None , **__lowercase : int , ) -> Optional[Any]: super().__init__(**_A ) SCREAMING_SNAKE_CASE__ : Optional[int] =patch_size SCREAMING_SNAKE_CASE__ : Dict =num_channels SCREAMING_SNAKE_CASE__ : str =embed_dim SCREAMING_SNAKE_CASE__ : str =depths SCREAMING_SNAKE_CASE__ : Dict =len(_A ) SCREAMING_SNAKE_CASE__ : Any =num_heads SCREAMING_SNAKE_CASE__ : Any =kernel_size SCREAMING_SNAKE_CASE__ : List[str] =dilations SCREAMING_SNAKE_CASE__ : Union[str, Any] =mlp_ratio SCREAMING_SNAKE_CASE__ : Optional[Any] =qkv_bias SCREAMING_SNAKE_CASE__ : List[Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Dict =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : List[Any] =drop_path_rate SCREAMING_SNAKE_CASE__ : Optional[Any] =hidden_act SCREAMING_SNAKE_CASE__ : Union[str, Any] =layer_norm_eps SCREAMING_SNAKE_CASE__ : List[Any] =initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ : List[str] =int(embed_dim * 2 ** (len(_A ) - 1) ) SCREAMING_SNAKE_CASE__ : Dict =layer_scale_init_value SCREAMING_SNAKE_CASE__ : int =['stem'] + [F"stage{idx}" for idx in range(1 , len(_A ) + 1 )] SCREAMING_SNAKE_CASE__ : Tuple =get_aligned_output_features_output_indices( out_features=_A , out_indices=_A , stage_names=self.stage_names )
711
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = """gpt_neox""" def __init__( self : List[Any] , __lowercase : Union[str, Any]=5_04_32 , __lowercase : int=61_44 , __lowercase : Tuple=44 , __lowercase : List[str]=64 , __lowercase : str=2_45_76 , __lowercase : Dict="gelu" , __lowercase : Tuple=0.25 , __lowercase : Tuple=1_00_00 , __lowercase : Tuple=0.0 , __lowercase : str=0.0 , __lowercase : List[Any]=0.1 , __lowercase : Dict=20_48 , __lowercase : Any=0.02 , __lowercase : Dict=1e-5 , __lowercase : List[Any]=True , __lowercase : str=0 , __lowercase : Optional[Any]=2 , __lowercase : Tuple=False , __lowercase : List[Any]=True , __lowercase : Optional[Any]=None , **__lowercase : Any , ) -> Dict: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase ) SCREAMING_SNAKE_CASE__ : Optional[int] =vocab_size SCREAMING_SNAKE_CASE__ : Union[str, Any] =max_position_embeddings SCREAMING_SNAKE_CASE__ : Any =hidden_size SCREAMING_SNAKE_CASE__ : str =num_hidden_layers SCREAMING_SNAKE_CASE__ : Any =num_attention_heads SCREAMING_SNAKE_CASE__ : Union[str, Any] =intermediate_size SCREAMING_SNAKE_CASE__ : Dict =hidden_act SCREAMING_SNAKE_CASE__ : str =rotary_pct SCREAMING_SNAKE_CASE__ : Optional[Any] =rotary_emb_base SCREAMING_SNAKE_CASE__ : List[Any] =attention_dropout SCREAMING_SNAKE_CASE__ : List[Any] =hidden_dropout SCREAMING_SNAKE_CASE__ : str =classifier_dropout SCREAMING_SNAKE_CASE__ : Any =initializer_range SCREAMING_SNAKE_CASE__ : Dict =layer_norm_eps SCREAMING_SNAKE_CASE__ : Any =use_cache SCREAMING_SNAKE_CASE__ : Tuple =tie_word_embeddings SCREAMING_SNAKE_CASE__ : Tuple =use_parallel_residual SCREAMING_SNAKE_CASE__ : Union[str, Any] =rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( '''The hidden size is not divisble by the number of attention heads! Make sure to update them!''' ) def __magic_name__ ( self : Optional[Any] ) -> Optional[Any]: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __lowercase ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"got {self.rope_scaling}" ) SCREAMING_SNAKE_CASE__ : int =self.rope_scaling.get('''type''' , __lowercase ) SCREAMING_SNAKE_CASE__ : Optional[int] =self.rope_scaling.get('''factor''' , __lowercase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(__lowercase , __lowercase ) or rope_scaling_factor <= 1.0: raise ValueError(F"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
665
0
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __snake_case = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase ( __snake_case , unittest.TestCase ): lowercase = XLMRobertaTokenizer lowercase = XLMRobertaTokenizerFast lowercase = True lowercase = True def lowerCamelCase_ ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase = XLMRobertaTokenizer(__magic_name__ , keep_accents=__magic_name__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = """<pad>""" UpperCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__magic_name__ ) , 1_0_0_2 ) def lowerCamelCase_ ( self : int ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = XLMRobertaTokenizer(__magic_name__ , keep_accents=__magic_name__ ) UpperCamelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__magic_name__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__magic_name__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) UpperCamelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __magic_name__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCamelCase = tokenizer.convert_tokens_to_ids(__magic_name__ ) self.assertListEqual( __magic_name__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCamelCase = tokenizer.convert_ids_to_tokens(__magic_name__ ) self.assertListEqual( __magic_name__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def lowerCamelCase_ ( self : int ): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCamelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__magic_name__ , **__magic_name__ ) UpperCamelCase = self.tokenizer_class.from_pretrained(__magic_name__ , **__magic_name__ ) UpperCamelCase = tempfile.mkdtemp() UpperCamelCase = tokenizer_r.save_pretrained(__magic_name__ ) UpperCamelCase = tokenizer_p.save_pretrained(__magic_name__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCamelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(__magic_name__ , __magic_name__ ) # Checks everything loads correctly in the same way UpperCamelCase = tokenizer_r.from_pretrained(__magic_name__ ) UpperCamelCase = tokenizer_p.from_pretrained(__magic_name__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__magic_name__ , __magic_name__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__magic_name__ ) # Save tokenizer rust, legacy_format=True UpperCamelCase = tempfile.mkdtemp() UpperCamelCase = tokenizer_r.save_pretrained(__magic_name__ , legacy_format=__magic_name__ ) UpperCamelCase = tokenizer_p.save_pretrained(__magic_name__ ) # Checks it save with the same files self.assertSequenceEqual(__magic_name__ , __magic_name__ ) # Checks everything loads correctly in the same way UpperCamelCase = tokenizer_r.from_pretrained(__magic_name__ ) UpperCamelCase = tokenizer_p.from_pretrained(__magic_name__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__magic_name__ , __magic_name__ ) ) shutil.rmtree(__magic_name__ ) # Save tokenizer rust, legacy_format=False UpperCamelCase = tempfile.mkdtemp() UpperCamelCase = tokenizer_r.save_pretrained(__magic_name__ , legacy_format=__magic_name__ ) UpperCamelCase = tokenizer_p.save_pretrained(__magic_name__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCamelCase = tokenizer_r.from_pretrained(__magic_name__ ) UpperCamelCase = tokenizer_p.from_pretrained(__magic_name__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__magic_name__ , __magic_name__ ) ) shutil.rmtree(__magic_name__ ) @cached_property def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__magic_name__ , f.name ) UpperCamelCase = XLMRobertaTokenizer(f.name , keep_accents=__magic_name__ ) UpperCamelCase = pickle.dumps(__magic_name__ ) pickle.loads(__magic_name__ ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" if not self.test_rust_tokenizer: return UpperCamelCase = self.get_tokenizer() UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = """I was born in 92000, and this is falsé.""" UpperCamelCase = tokenizer.tokenize(__magic_name__ ) UpperCamelCase = rust_tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) UpperCamelCase = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) UpperCamelCase = rust_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = tokenizer.encode(__magic_name__ ) UpperCamelCase = rust_tokenizer.encode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) @slow def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = """Hello World!""" UpperCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__magic_name__ , self.big_tokenizer.encode(__magic_name__ ) ) @slow def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__magic_name__ , self.big_tokenizer.encode(__magic_name__ ) ) @slow def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = {"""input_ids""": [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
386
import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = 0 @slow def lowerCamelCase_ ( self : Dict ): """simple docstring""" for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(__magic_name__ ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) self.assertIsInstance(__magic_name__ , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(__magic_name__ ) , 0 ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 1_2 ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 2_0 ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = AutoConfig.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) # Check that tokenizer_type ≠ model_type UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , config=__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 1_2 ) def lowerCamelCase_ ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(__magic_name__ , """vocab.txt""" ) ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""bert""" , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(__magic_name__ , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(__magic_name__ , """merges.txt""" ) ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""gpt2""" , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) @require_tokenizers def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(__magic_name__ , """vocab.txt""" ) ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""bert""" ) self.assertIsInstance(__magic_name__ , __magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(__magic_name__ , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(__magic_name__ , """merges.txt""" ) ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""gpt2""" ) self.assertIsInstance(__magic_name__ , __magic_name__ ) def lowerCamelCase_ ( self : Any ): """simple docstring""" with pytest.raises(__magic_name__ ): AutoTokenizer.from_pretrained("""./""" , tokenizer_type="""xxx""" ) @require_tokenizers def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: UpperCamelCase = tokenizer_class.from_pretrained("""wietsedv/bert-base-dutch-cased""" ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) if isinstance(__magic_name__ , __magic_name__ ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , __magic_name__ ) else: self.assertEqual(tokenizer.do_lower_case , __magic_name__ ) self.assertEqual(tokenizer.model_max_length , 5_1_2 ) @require_tokenizers def lowerCamelCase_ ( self : Dict ): """simple docstring""" for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( __magic_name__ , """julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier""" , ): UpperCamelCase = tokenizer_class.from_pretrained("""julien-c/herlolip-not-exists""" ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = TOKENIZER_MAPPING.values() UpperCamelCase = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(__magic_name__ ) @require_tokenizers def lowerCamelCase_ ( self : Any ): """simple docstring""" self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=__magic_name__ ) , __magic_name__ ) self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" ) , __magic_name__ ) @require_tokenizers def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained("""distilbert-base-uncased""" , do_lower_case=__magic_name__ ) UpperCamelCase = """Hello, world. How are you?""" UpperCamelCase = tokenizer.tokenize(__magic_name__ ) self.assertEqual("""[UNK]""" , tokens[0] ) UpperCamelCase = AutoTokenizer.from_pretrained("""microsoft/mpnet-base""" , do_lower_case=__magic_name__ ) UpperCamelCase = tokenizer.tokenize(__magic_name__ ) self.assertEqual("""[UNK]""" , tokens[0] ) @require_tokenizers def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained("""robot-test/dummy-tokenizer-fast-with-model-config""" ) self.assertEqual(type(__magic_name__ ) , __magic_name__ ) self.assertEqual(tokenizer.model_max_length , 5_1_2 ) self.assertEqual(tokenizer.vocab_size , 3_0_0_0_0 ) self.assertEqual(tokenizer.unk_token , """[UNK]""" ) self.assertEqual(tokenizer.padding_side , """right""" ) self.assertEqual(tokenizer.truncation_side , """right""" ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 1_2 ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained("""ctrl""" ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(__magic_name__ , __magic_name__ ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = get_tokenizer_config("""bert-base-cased""" ) UpperCamelCase = config.pop("""_commit_hash""" , __magic_name__ ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(__magic_name__ , {"""do_lower_case""": False} ) # This model does not have a tokenizer_config so we get back an empty dict. UpperCamelCase = get_tokenizer_config(__magic_name__ ) self.assertDictEqual(__magic_name__ , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = get_tokenizer_config(__magic_name__ ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config["""tokenizer_class"""] , """BertTokenizer""" ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" try: AutoConfig.register("""custom""" , __magic_name__ ) AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__magic_name__ ): AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) UpperCamelCase = CustomTokenizer.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" try: AutoConfig.register("""custom""" , __magic_name__ ) # Can register in two steps AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( __magic_name__ , slow_tokenizer_class=__magic_name__ , fast_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__magic_name__ ): AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: UpperCamelCase = BertTokenizerFast.from_pretrained(__magic_name__ ) bert_tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = CustomTokenizerFast.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" with self.assertRaises(__magic_name__ ): UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__magic_name__ ): UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , trust_remote_code=__magic_name__ ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) @require_tokenizers def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" class UpperCAmelCase ( __snake_case ): lowercase = False class UpperCAmelCase ( __snake_case ): lowercase = NewTokenizer lowercase = False try: AutoConfig.register("""custom""" , __magic_name__ ) AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) # If remote code is not set, the default is to use local UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertTrue(tokenizer.special_attribute_present ) UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version UpperCamelCase = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) def lowerCamelCase_ ( self : Any ): """simple docstring""" with self.assertRaisesRegex( __magic_name__ , """bert-base is not a local folder and is not a valid model identifier""" ): UpperCamelCase = AutoTokenizer.from_pretrained("""bert-base""" ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( __magic_name__ , R"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): UpperCamelCase = AutoTokenizer.from_pretrained(__magic_name__ , revision="""aaaaaa""" ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) with RequestCounter() as counter: UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
386
1
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def UpperCamelCase__ ( __magic_name__ : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case__ : int = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", """decoder.output_projection.weight""", ] for k in ignore_keys: state_dict.pop(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( __magic_name__ : Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case__ , snake_case__ : str = emb.weight.shape snake_case__ : List[str] = nn.Linear(__magic_name__ , __magic_name__ , bias=__magic_name__ ) snake_case__ : Tuple = emb.weight.data return lin_layer def UpperCamelCase__ ( __magic_name__ : List[Any] , __magic_name__ : List[Any]="facebook/mbart-large-en-ro" , __magic_name__ : int=False , __magic_name__ : Tuple=False ) -> str: '''simple docstring''' snake_case__ : Optional[Any] = torch.load(__magic_name__ , map_location="""cpu""" )["""model"""] remove_ignore_keys_(__magic_name__ ) snake_case__ : Dict = state_dict["""encoder.embed_tokens.weight"""].shape[0] snake_case__ : Union[str, Any] = MBartConfig.from_pretrained(__magic_name__ , vocab_size=__magic_name__ ) if mbart_aa and finetuned: snake_case__ : str = """relu""" snake_case__ : int = state_dict["""decoder.embed_tokens.weight"""] snake_case__ : int = MBartForConditionalGeneration(__magic_name__ ) model.model.load_state_dict(__magic_name__ ) if finetuned: snake_case__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": A_ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "fairseq_path", type=str, help="bart.large, bart.large.cnn or a path to a model.pt on local filesystem." ) parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--hf_config", default="facebook/mbart-large-cc25", type=str, help="Which huggingface architecture to use: mbart-large", ) parser.add_argument("--mbart_50", action="store_true", help="whether the model is mMART-50 checkpoint") parser.add_argument("--finetuned", action="store_true", help="whether the model is a fine-tuned checkpoint") A_ : Any = parser.parse_args() A_ : Tuple = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
711
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __snake_case ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = StableDiffusionInstructPixaPixPipeline lowerCamelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width''', '''cross_attention_kwargs'''} lowerCamelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def __UpperCamelCase ( self ): torch.manual_seed(0 ) snake_case__ : Optional[Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , ) snake_case__ : Any = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE ) torch.manual_seed(0 ) snake_case__ : List[str] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case__ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) snake_case__ : Optional[Any] = CLIPTextModel(__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case__ : Tuple = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ): snake_case__ : List[str] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case__ : Any = Image.fromarray(np.uinta(__SCREAMING_SNAKE_CASE ) ).convert("""RGB""" ) if str(__SCREAMING_SNAKE_CASE ).startswith("""mps""" ): snake_case__ : List[str] = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: snake_case__ : List[Any] = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """image_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def __UpperCamelCase ( self ): snake_case__ : Dict = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case__ : str = self.get_dummy_components() snake_case__ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline(**__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) snake_case__ : int = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = sd_pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : Tuple = np.array([0.7526, 0.3750, 0.4547, 0.6117, 0.5866, 0.5016, 0.4327, 0.5642, 0.4815] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : Tuple = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case__ : Tuple = self.get_dummy_components() snake_case__ : Optional[int] = StableDiffusionInstructPixaPixPipeline(**__SCREAMING_SNAKE_CASE ) snake_case__ : Union[str, Any] = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) snake_case__ : Any = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) snake_case__ : Any = """french fries""" snake_case__ : Union[str, Any] = sd_pipe(**__SCREAMING_SNAKE_CASE , negative_prompt=__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = output.images snake_case__ : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : str = np.array([0.7511, 0.3642, 0.4553, 0.6236, 0.5797, 0.5013, 0.4343, 0.5611, 0.4831] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case__ : str = self.get_dummy_components() snake_case__ : List[Any] = StableDiffusionInstructPixaPixPipeline(**__SCREAMING_SNAKE_CASE ) snake_case__ : int = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = [inputs["""prompt"""]] * 2 snake_case__ : List[str] = np.array(inputs["""image"""] ).astype(np.floataa ) / 255.0 snake_case__ : int = torch.from_numpy(__SCREAMING_SNAKE_CASE ).unsqueeze(0 ).to(__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = image / 2 + 0.5 snake_case__ : str = image.permute(0 , 3 , 1 , 2 ) snake_case__ : List[str] = image.repeat(2 , 1 , 1 , 1 ) snake_case__ : Optional[int] = sd_pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : Union[str, Any] = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) snake_case__ : Optional[Any] = np.array([0.5812, 0.5748, 0.5222, 0.5908, 0.5695, 0.7174, 0.6804, 0.5523, 0.5579] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : List[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case__ : Optional[Any] = self.get_dummy_components() snake_case__ : int = EulerAncestralDiscreteScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" ) snake_case__ : List[str] = StableDiffusionInstructPixaPixPipeline(**__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) snake_case__ : int = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[Any] = sd_pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : List[Any] = image[0, -3:, -3:, -1] snake_case__ : List[str] = [round(__SCREAMING_SNAKE_CASE , 4 ) for x in image_slice.flatten().tolist()] print(""",""".join([str(__SCREAMING_SNAKE_CASE ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : List[str] = np.array([0.7417, 0.3842, 0.4732, 0.5776, 0.5891, 0.5139, 0.4052, 0.5673, 0.4986] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def __UpperCamelCase ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def __UpperCamelCase ( self ): snake_case__ : Any = self.get_dummy_components() snake_case__ : Any = StableDiffusionInstructPixaPixPipeline(**__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = VaeImageProcessor(do_resize=__SCREAMING_SNAKE_CASE , do_normalize=__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) snake_case__ : Union[str, Any] = pipe(**self.get_dummy_inputs_by_type(__SCREAMING_SNAKE_CASE , input_image_type="""pt""" ) )[0] snake_case__ : Optional[int] = components["""vae"""] snake_case__ : int = self.get_dummy_inputs_by_type(__SCREAMING_SNAKE_CASE , input_image_type="""pt""" ) for image_param in self.image_latents_params: if image_param in inputs.keys(): snake_case__ : Optional[Any] = vae.encode(inputs[image_param] ).latent_dist.mode() snake_case__ : Union[str, Any] = pipe(**__SCREAMING_SNAKE_CASE )[0] snake_case__ : Union[str, Any] = np.abs(out - out_latents_inputs ).max() self.assertLess(__SCREAMING_SNAKE_CASE , 1e-4 , """passing latents as image input generate different result from passing image""" ) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE=0 ): snake_case__ : Optional[Any] = torch.manual_seed(__SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = load_image( """https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg""" ) snake_case__ : Tuple = { """prompt""": """turn him into a cyborg""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """image_guidance_scale""": 1.0, """output_type""": """numpy""", } return inputs def __UpperCamelCase ( self ): snake_case__ : List[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() snake_case__ : Dict = self.get_inputs() snake_case__ : Tuple = pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : Dict = np.array([0.5902, 0.6015, 0.6027, 0.5983, 0.6092, 0.6061, 0.5765, 0.5785, 0.5555] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__SCREAMING_SNAKE_CASE ) snake_case__ : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() snake_case__ : Tuple = self.get_inputs() snake_case__ : Union[str, Any] = pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : int = np.array([0.6578, 0.6817, 0.6972, 0.6761, 0.6856, 0.6916, 0.6428, 0.6516, 0.6301] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : Any = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() snake_case__ : Dict = self.get_inputs() snake_case__ : Tuple = pipe(**__SCREAMING_SNAKE_CASE ).images snake_case__ : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : Optional[int] = np.array([0.3828, 0.3834, 0.3818, 0.3792, 0.3865, 0.3752, 0.3792, 0.3847, 0.3753] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def __UpperCamelCase ( self ): snake_case__ : List[str] = 0 def callback_fn(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> None: snake_case__ : Tuple = True nonlocal number_of_steps number_of_steps += 1 if step == 1: snake_case__ : Optional[int] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) snake_case__ : Optional[Any] = latents[0, -3:, -3:, -1] snake_case__ : Optional[int] = np.array([-0.2463, -0.4644, -0.9756, 1.5176, 1.4414, 0.7866, 0.9897, 0.8521, 0.7983] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: snake_case__ : Optional[int] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) snake_case__ : List[str] = latents[0, -3:, -3:, -1] snake_case__ : int = np.array([-0.2644, -0.4626, -0.9653, 1.5176, 1.4551, 0.7686, 0.9805, 0.8452, 0.8115] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 snake_case__ : List[str] = False snake_case__ : str = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa ) snake_case__ : Tuple = pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() snake_case__ : Tuple = self.get_inputs() pipe(**__SCREAMING_SNAKE_CASE , callback=__SCREAMING_SNAKE_CASE , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def __UpperCamelCase ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case__ : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa ) snake_case__ : str = pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case__ : str = self.get_inputs() snake_case__ : Any = pipe(**__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def __UpperCamelCase ( self ): snake_case__ : Union[str, Any] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 snake_case__ : Tuple = inputs["""image"""].resize((5_0_4, 5_0_4) ) snake_case__ : Dict = """timbrooks/instruct-pix2pix""" snake_case__ : Optional[int] = StableDiffusionInstructPixaPixPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() snake_case__ : Tuple = pipe(**__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = output.images[0] snake_case__ : Union[str, Any] = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) snake_case__ : int = np.array([0.2726, 0.2529, 0.2664, 0.2655, 0.2641, 0.2642, 0.2591, 0.2649, 0.2590] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
419
0
'''simple docstring''' import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed a = { """distilbert""": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), """roberta""": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), """bert""": (BertConfig, BertForMaskedLM, BertTokenizer), """gpt2""": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def __magic_name__ ( __UpperCAmelCase ) -> int: '''simple docstring''' assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' if args.student_type == "roberta": __SCREAMING_SNAKE_CASE = False elif args.student_type == "gpt2": __SCREAMING_SNAKE_CASE = False def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' if args.student_type == "roberta": __SCREAMING_SNAKE_CASE = False def __magic_name__ ( ) -> Union[str, Any]: '''simple docstring''' __SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Training""" ) parser.add_argument("""--force""" , action="""store_true""" , help="""Overwrite dump_path if it already exists.""" ) parser.add_argument( """--dump_path""" , type=__snake_case , required=__snake_case , help="""The output directory (log, checkpoints, parameters, etc.)""" ) parser.add_argument( """--data_file""" , type=__snake_case , required=__snake_case , help="""The binarized file (tokenized + tokens_to_ids) and grouped by sequence.""" , ) parser.add_argument( """--student_type""" , type=__snake_case , choices=["""distilbert""", """roberta""", """gpt2"""] , required=__snake_case , help="""The student type (DistilBERT, RoBERTa).""" , ) parser.add_argument("""--student_config""" , type=__snake_case , required=__snake_case , help="""Path to the student configuration.""" ) parser.add_argument( """--student_pretrained_weights""" , default=__snake_case , type=__snake_case , help="""Load student initialization checkpoint.""" ) parser.add_argument( """--teacher_type""" , choices=["""bert""", """roberta""", """gpt2"""] , required=__snake_case , help="""Teacher type (BERT, RoBERTa).""" ) parser.add_argument("""--teacher_name""" , type=__snake_case , required=__snake_case , help="""The teacher model.""" ) parser.add_argument("""--temperature""" , default=2.0 , type=__snake_case , help="""Temperature for the softmax temperature.""" ) parser.add_argument( """--alpha_ce""" , default=0.5 , type=__snake_case , help="""Linear weight for the distillation loss. Must be >=0.""" ) parser.add_argument( """--alpha_mlm""" , default=0.0 , type=__snake_case , help="""Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.""" , ) parser.add_argument("""--alpha_clm""" , default=0.5 , type=__snake_case , help="""Linear weight for the CLM loss. Must be >=0.""" ) parser.add_argument("""--alpha_mse""" , default=0.0 , type=__snake_case , help="""Linear weight of the MSE loss. Must be >=0.""" ) parser.add_argument( """--alpha_cos""" , default=0.0 , type=__snake_case , help="""Linear weight of the cosine embedding loss. Must be >=0.""" ) parser.add_argument( """--mlm""" , action="""store_true""" , help="""The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.""" ) parser.add_argument( """--mlm_mask_prop""" , default=0.1_5 , type=__snake_case , help="""Proportion of tokens for which we need to make a prediction.""" , ) parser.add_argument("""--word_mask""" , default=0.8 , type=__snake_case , help="""Proportion of tokens to mask out.""" ) parser.add_argument("""--word_keep""" , default=0.1 , type=__snake_case , help="""Proportion of tokens to keep.""" ) parser.add_argument("""--word_rand""" , default=0.1 , type=__snake_case , help="""Proportion of tokens to randomly replace.""" ) parser.add_argument( """--mlm_smoothing""" , default=0.7 , type=__snake_case , help="""Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).""" , ) parser.add_argument("""--token_counts""" , type=__snake_case , help="""The token counts in the data_file for MLM.""" ) parser.add_argument( """--restrict_ce_to_mask""" , action="""store_true""" , help="""If true, compute the distillation loss only the [MLM] prediction distribution.""" , ) parser.add_argument( """--freeze_pos_embs""" , action="""store_true""" , help="""Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.""" , ) parser.add_argument( """--freeze_token_type_embds""" , action="""store_true""" , help="""Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.""" , ) parser.add_argument("""--n_epoch""" , type=__snake_case , default=3 , help="""Number of pass on the whole dataset.""" ) parser.add_argument("""--batch_size""" , type=__snake_case , default=5 , help="""Batch size (for each process).""" ) parser.add_argument( """--group_by_size""" , action="""store_false""" , help="""If true, group sequences that have similar length into the same batch. Default is true.""" , ) parser.add_argument( """--gradient_accumulation_steps""" , type=__snake_case , default=50 , help="""Gradient accumulation for larger training batches.""" , ) parser.add_argument("""--warmup_prop""" , default=0.0_5 , type=__snake_case , help="""Linear warmup proportion.""" ) parser.add_argument("""--weight_decay""" , default=0.0 , type=__snake_case , help="""Weight decay if we apply some.""" ) parser.add_argument("""--learning_rate""" , default=5e-4 , type=__snake_case , help="""The initial learning rate for Adam.""" ) parser.add_argument("""--adam_epsilon""" , default=1e-6 , type=__snake_case , help="""Epsilon for Adam optimizer.""" ) parser.add_argument("""--max_grad_norm""" , default=5.0 , type=__snake_case , help="""Max gradient norm.""" ) parser.add_argument("""--initializer_range""" , default=0.0_2 , type=__snake_case , help="""Random initialization range.""" ) parser.add_argument( """--fp16""" , action="""store_true""" , help="""Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit""" , ) parser.add_argument( """--fp16_opt_level""" , type=__snake_case , default="""O1""" , help=( """For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].""" """See details at https://nvidia.github.io/apex/amp.html""" ) , ) parser.add_argument("""--n_gpu""" , type=__snake_case , default=1 , help="""Number of GPUs in the node.""" ) parser.add_argument("""--local_rank""" , type=__snake_case , default=-1 , help="""Distributed training - Local rank""" ) parser.add_argument("""--seed""" , type=__snake_case , default=56 , help="""Random seed""" ) parser.add_argument("""--log_interval""" , type=__snake_case , default=500 , help="""Tensorboard logging interval.""" ) parser.add_argument("""--checkpoint_interval""" , type=__snake_case , default=4000 , help="""Checkpoint interval.""" ) __SCREAMING_SNAKE_CASE = parser.parse_args() sanity_checks(__snake_case ) # ARGS # init_gpu_params(__snake_case ) set_seed(__snake_case ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" """ itUse `--force` if you want to overwrite it""" ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(f"""Param: {args}""" ) with open(os.path.join(args.dump_path , """parameters.json""" ) , """w""" ) as f: json.dump(vars(__snake_case ) , __snake_case , indent=4 ) git_log(args.dump_path ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = MODEL_CLASSES[args.student_type] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = MODEL_CLASSES[args.teacher_type] # TOKENIZER # __SCREAMING_SNAKE_CASE = teacher_tokenizer_class.from_pretrained(args.teacher_name ) __SCREAMING_SNAKE_CASE = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): __SCREAMING_SNAKE_CASE = tokenizer.all_special_tokens.index(__snake_case ) __SCREAMING_SNAKE_CASE = tokenizer.all_special_ids[idx] logger.info(f"""Special tokens {special_tok_ids}""" ) __SCREAMING_SNAKE_CASE = special_tok_ids __SCREAMING_SNAKE_CASE = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"""Loading data from {args.data_file}""" ) with open(args.data_file , """rb""" ) as fp: __SCREAMING_SNAKE_CASE = pickle.load(__snake_case ) if args.mlm: logger.info(f"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts , """rb""" ) as fp: __SCREAMING_SNAKE_CASE = pickle.load(__snake_case ) __SCREAMING_SNAKE_CASE = np.maximum(__snake_case , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): __SCREAMING_SNAKE_CASE = 0.0 # do not predict special tokens __SCREAMING_SNAKE_CASE = torch.from_numpy(__snake_case ) else: __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = LmSeqsDataset(params=__snake_case , data=__snake_case ) logger.info("""Data loader created.""" ) # STUDENT # logger.info(f"""Loading student config from {args.student_config}""" ) __SCREAMING_SNAKE_CASE = student_config_class.from_pretrained(args.student_config ) __SCREAMING_SNAKE_CASE = True if args.student_pretrained_weights is not None: logger.info(f"""Loading pretrained weights from {args.student_pretrained_weights}""" ) __SCREAMING_SNAKE_CASE = student_model_class.from_pretrained(args.student_pretrained_weights , config=__snake_case ) else: __SCREAMING_SNAKE_CASE = student_model_class(__snake_case ) if args.n_gpu > 0: student.to(f"""cuda:{args.local_rank}""" ) logger.info("""Student loaded.""" ) # TEACHER # __SCREAMING_SNAKE_CASE = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=__snake_case ) if args.n_gpu > 0: teacher.to(f"""cuda:{args.local_rank}""" ) logger.info(f"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__snake_case , __snake_case ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__snake_case , __snake_case ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() __SCREAMING_SNAKE_CASE = Distiller( params=__snake_case , dataset=__snake_case , token_probs=__snake_case , student=__snake_case , teacher=__snake_case ) distiller.train() logger.info("""Let\'s go get some drinks.""" ) if __name__ == "__main__": main()
109
"""simple docstring""" def A ( __snake_case: str ) -> list: """simple docstring""" return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(__snake_case ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__("""doctest""").testmod()
545
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase = { """configuration_llama""": ["""LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LlamaConfig"""], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["""LlamaTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["""LlamaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """LlamaForCausalLM""", """LlamaModel""", """LlamaPreTrainedModel""", """LlamaForSequenceClassification""", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys _lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
711
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): _lowerCamelCase = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right _lowerCamelCase = 128022 _lowerCamelCase = 128028 @require_sentencepiece class _SCREAMING_SNAKE_CASE (UpperCamelCase , unittest.TestCase ): lowerCAmelCase = MaMaaaTokenizer lowerCAmelCase = False lowerCAmelCase = False lowerCAmelCase = True def __snake_case ( self : Union[str, Any] )->Optional[Any]: super().setUp() __SCREAMING_SNAKE_CASE : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] __SCREAMING_SNAKE_CASE : List[str] = dict(zip(UpperCamelCase , range(len(UpperCamelCase ) ) ) ) __SCREAMING_SNAKE_CASE : int = Path(self.tmpdirname ) save_json(UpperCamelCase , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(UpperCamelCase , save_dir / VOCAB_FILES_NAMES["spm_file"] ) __SCREAMING_SNAKE_CASE : Any = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[int] , **UpperCamelCase : Any )->Dict: return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase ) def __snake_case ( self : Dict , UpperCamelCase : List[str] )->int: return ( "This is a test", "This is a test", ) def __snake_case ( self : str )->Tuple: __SCREAMING_SNAKE_CASE : List[Any] = "</s>" __SCREAMING_SNAKE_CASE : Optional[int] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase ) , UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase ) , UpperCamelCase ) def __snake_case ( self : Tuple )->Optional[Any]: __SCREAMING_SNAKE_CASE : Optional[Any] = self.get_tokenizer() __SCREAMING_SNAKE_CASE : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(UpperCamelCase ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def __snake_case ( self : Dict )->Dict: pass def __snake_case ( self : Union[str, Any] )->Union[str, Any]: __SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_tokenizer() __SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(UpperCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase ) , [2, 3, 4, 5, 6] , ) __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(UpperCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) __SCREAMING_SNAKE_CASE : Any = tokenizer.convert_tokens_to_string(UpperCamelCase ) self.assertEqual(UpperCamelCase , "This is a test" ) @slow def __snake_case ( self : Any )->Union[str, Any]: # fmt: off __SCREAMING_SNAKE_CASE : Tuple = {"input_ids": [[1_2_8_0_2_2, 1_1_0_1_0_8, 3_9_7, 1_1, 3_8_2_7_2, 2_2_4_7, 1_2_4_8_1_1, 2_8_5, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 3_9_5_3_4, 4_4_2_8, 3_9_7, 1_0_1_9, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 4_1_3_3_7, 1_6_7_8_6, 2_4_1, 7, 2_0_2_1_4, 1_7, 1_2_5_6_9_0, 1_0_3_9_8, 7, 4_4_3_7_8, 5_8_0_6_9, 6_8_3_4_2, 7_7_9_8, 7_3_4_3, 1_1, 2_9_9, 3_3_3_1_0, 4, 1_5_8, 3_7_3_5_0, 9_4_0_7_7, 4_5_6_9, 2_9_9, 3_3_3_1_0, 9_0, 4, 5_2_8_4_0, 2_9_0, 4, 3_1_2_7_0, 1_1_2, 2_9_9, 6_8_2, 4, 5_2_8_4_0, 3_9_9_5_3, 1_4_0_7_9, 1_9_3, 5_2_5_1_9, 9_0_8_9_4, 1_7_8_9_4, 1_2_0_6_9_7, 1_1, 4_0_4_4_5, 5_5_1, 1_7, 1_0_1_9, 5_2_5_1_9, 9_0_8_9_4, 1_7_7_5_6, 9_6_3, 1_1, 4_0_4_4_5, 4_8_0, 1_7, 9_7_9_2, 1_1_2_0, 5_1_7_3, 1_3_9_3, 6_2_4_0, 1_6_7_8_6, 2_4_1, 1_2_0_9_9_6, 2_8, 1_2_4_5, 1_3_9_3, 1_1_8_2_4_0, 1_1_1_2_3, 1_0_1_9, 9_3_6_1_2, 2_6_9_1, 1_0_6_1_8, 9_8_0_5_8, 1_2_0_4_0_9, 1_9_2_8, 2_7_9, 4, 4_0_6_8_3, 3_6_7, 1_7_8, 2_0_7, 1_0_1_9, 1_0_3, 1_0_3_1_2_1, 5_0_6, 6_5_2_9_6, 5, 2], [1_2_8_0_2_2, 2_1_2_1_7, 3_6_7, 1_1_7, 1_2_5_4_5_0, 1_2_8, 7_1_9, 7, 7_3_0_8, 4_0, 9_3_6_1_2, 1_2_6_6_9, 1_1_1_6, 1_6_7_0_4, 7_1, 1_7_7_8_5, 3_6_9_9, 1_5_5_9_2, 3_5, 1_4_4, 9_5_8_4, 2_4_1, 1_1_9_4_3, 7_1_3, 9_5_0, 7_9_9, 2_2_4_7, 8_8_4_2_7, 1_5_0, 1_4_9, 1_1_8_8_1_3, 1_2_0_7_0_6, 1_0_1_9, 1_0_6_9_0_6, 8_1_5_1_8, 2_8, 1_2_2_4, 2_2_7_9_9, 3_9_7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1_2_8_0_2_2, 1_6_5_8, 1_2_3_3_1_1, 5_1_5_5, 5_5_7_8, 4_7_2_2, 2_7_9, 1_4_9_4_7, 2_3_6_6, 1_1_2_0, 1_1_9_7, 1_4, 1_3_4_8, 9_2_3_2, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class _SCREAMING_SNAKE_CASE (unittest.TestCase ): lowerCAmelCase = """facebook/m2m100_418M""" lowerCAmelCase = [ """In my opinion, there are two levels of response from the French government.""", """NSA Affair Emphasizes Complete Lack of Debate on Intelligence""", ] lowerCAmelCase = [ """Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.""", """L'affaire NSA souligne l'absence totale de débat sur le renseignement""", ] # fmt: off lowerCAmelCase = [EN_CODE, 593, 1949, 11_5781, 4, 7_1586, 4234, 6_0633, 12_6233, 432, 12_3808, 1_5592, 1197, 11_7132, 12_0618, 5, 2] @classmethod def __snake_case ( cls : List[Any] )->Union[str, Any]: __SCREAMING_SNAKE_CASE : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) __SCREAMING_SNAKE_CASE : Optional[Any] = 1 return cls def __snake_case ( self : Dict )->Any: self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 1_2_8_0_0_6 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 1_2_8_0_2_2 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 1_2_8_0_7_6 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 1_2_8_0_6_3 ) def __snake_case ( self : Any )->Union[str, Any]: __SCREAMING_SNAKE_CASE : str = self.tokenizer.get_vocab() self.assertEqual(len(UpperCamelCase ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , UpperCamelCase ) def __snake_case ( self : List[Any] )->str: __SCREAMING_SNAKE_CASE : Union[str, Any] = "en" __SCREAMING_SNAKE_CASE : Union[str, Any] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase ) def __snake_case ( self : Union[str, Any] )->List[Any]: self.assertIn(UpperCamelCase , self.tokenizer.all_special_ids ) # fmt: off __SCREAMING_SNAKE_CASE : Dict = [FR_CODE, 5_3_6_4, 8_2, 8_6_4_2, 4, 2_9_4, 4_7, 8, 1_4_0_2_8, 1_3_6, 3_2_8_6, 9_7_0_6, 6, 9_0_7_9_7, 6, 1_4_4_0_1_2, 1_6_2, 8_8_1_2_8, 3_0_0_6_1, 5, 2] # fmt: on __SCREAMING_SNAKE_CASE : int = self.tokenizer.decode(UpperCamelCase , skip_special_tokens=UpperCamelCase ) __SCREAMING_SNAKE_CASE : Optional[int] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase ) self.assertEqual(UpperCamelCase , UpperCamelCase ) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase ) def __snake_case ( self : Any )->List[Any]: __SCREAMING_SNAKE_CASE : int = tempfile.mkdtemp() __SCREAMING_SNAKE_CASE : List[Any] = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(UpperCamelCase ) __SCREAMING_SNAKE_CASE : Any = MaMaaaTokenizer.from_pretrained(UpperCamelCase ) self.assertDictEqual(new_tok.lang_token_to_id , UpperCamelCase ) @require_torch def __snake_case ( self : Any )->int: __SCREAMING_SNAKE_CASE : List[str] = "en" __SCREAMING_SNAKE_CASE : Dict = "fr" __SCREAMING_SNAKE_CASE : str = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase , return_tensors="pt" ) __SCREAMING_SNAKE_CASE : Optional[int] = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: __SCREAMING_SNAKE_CASE : Union[str, Any] = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def __snake_case ( self : str )->int: __SCREAMING_SNAKE_CASE : Optional[Any] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) __SCREAMING_SNAKE_CASE : Optional[int] = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def __snake_case ( self : Optional[Any] )->List[Any]: __SCREAMING_SNAKE_CASE : Union[str, Any] = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) __SCREAMING_SNAKE_CASE : Optional[int] = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def __snake_case ( self : Tuple )->Optional[int]: __SCREAMING_SNAKE_CASE : Any = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(UpperCamelCase ) , { # en_XX, A, test, EOS "input_ids": [[1_2_8_0_2_2, 5_8, 4_1_8_3, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 1_2_8_0_0_6, } , )
447
0
import unittest import numpy as np def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = None , ) -> np.ndarray: UpperCamelCase_: str = np.shape(UpperCAmelCase__ ) UpperCamelCase_: str = np.shape(UpperCAmelCase__ ) UpperCamelCase_: List[Any] = np.shape(UpperCAmelCase__ ) if shape_a[0] != shape_b[0]: UpperCamelCase_: Any = ( 'Expected the same number of rows for A and B. ' F'''Instead found A of size {shape_a} and B of size {shape_b}''' ) raise ValueError(UpperCAmelCase__ ) if shape_b[1] != shape_c[1]: UpperCamelCase_: int = ( 'Expected the same number of columns for B and C. ' F'''Instead found B of size {shape_b} and C of size {shape_c}''' ) raise ValueError(UpperCAmelCase__ ) UpperCamelCase_: Dict = pseudo_inv if a_inv is None: try: UpperCamelCase_: Optional[Any] = np.linalg.inv(UpperCAmelCase__ ) except np.linalg.LinAlgError: raise ValueError( 'Input matrix A is not invertible. Cannot compute Schur complement.' ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase( unittest.TestCase ): """simple docstring""" def _a ( self ): UpperCamelCase_: Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) UpperCamelCase_: Dict = np.array([[0, 3], [3, 0], [2, 3]] ) UpperCamelCase_: Tuple = np.array([[2, 1], [6, 3]] ) UpperCamelCase_: Tuple = schur_complement(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) UpperCamelCase_: Optional[Any] = np.block([[a, b], [b.T, c]] ) UpperCamelCase_: List[str] = np.linalg.det(_lowerCamelCase ) UpperCamelCase_: List[str] = np.linalg.det(_lowerCamelCase ) UpperCamelCase_: Dict = np.linalg.det(_lowerCamelCase ) self.assertAlmostEqual(_lowerCamelCase , det_a * det_s ) def _a ( self ): UpperCamelCase_: int = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) UpperCamelCase_: List[str] = np.array([[0, 3], [3, 0], [2, 3]] ) UpperCamelCase_: List[str] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(_lowerCamelCase ): schur_complement(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a ( self ): UpperCamelCase_: List[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) UpperCamelCase_: str = np.array([[0, 3], [3, 0], [2, 3]] ) UpperCamelCase_: List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(_lowerCamelCase ): schur_complement(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
57
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _UpperCAmelCase ( a ,unittest.TestCase ): '''simple docstring''' a__ =CodeGenTokenizer a__ =CodeGenTokenizerFast a__ =True a__ ={'''add_prefix_space''': True} a__ =False def __lowerCAmelCase ( self ) -> Optional[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCAmelCase : List[str] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] _UpperCAmelCase : str = dict(zip(A , range(len(A ) ) ) ) _UpperCAmelCase : Dict = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] _UpperCAmelCase : Union[str, Any] = {'''unk_token''': '''<unk>'''} _UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCAmelCase : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A ) ) def __lowerCAmelCase ( self , **A ) -> Tuple: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A ) def __lowerCAmelCase ( self , **A ) -> List[Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A ) def __lowerCAmelCase ( self , A ) -> Dict: _UpperCAmelCase : List[Any] = '''lower newer''' _UpperCAmelCase : Dict = '''lower newer''' return input_text, output_text def __lowerCAmelCase ( self ) -> Tuple: _UpperCAmelCase : int = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _UpperCAmelCase : Optional[int] = '''lower newer''' _UpperCAmelCase : Tuple = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] _UpperCAmelCase : List[Any] = tokenizer.tokenize(A , add_prefix_space=A ) self.assertListEqual(A , A ) _UpperCAmelCase : Dict = tokens + [tokenizer.unk_token] _UpperCAmelCase : Dict = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , A ) def __lowerCAmelCase ( self ) -> List[Any]: if not self.test_rust_tokenizer: return _UpperCAmelCase : Optional[int] = self.get_tokenizer() _UpperCAmelCase : Optional[int] = self.get_rust_tokenizer(add_prefix_space=A ) _UpperCAmelCase : Dict = '''lower newer''' # Testing tokenization _UpperCAmelCase : Optional[Any] = tokenizer.tokenize(A , add_prefix_space=A ) _UpperCAmelCase : Any = rust_tokenizer.tokenize(A ) self.assertListEqual(A , A ) # Testing conversion to ids without special tokens _UpperCAmelCase : Dict = tokenizer.encode(A , add_special_tokens=A , add_prefix_space=A ) _UpperCAmelCase : List[str] = rust_tokenizer.encode(A , add_special_tokens=A ) self.assertListEqual(A , A ) # Testing conversion to ids with special tokens _UpperCAmelCase : List[Any] = self.get_rust_tokenizer(add_prefix_space=A ) _UpperCAmelCase : List[str] = tokenizer.encode(A , add_prefix_space=A ) _UpperCAmelCase : Tuple = rust_tokenizer.encode(A ) self.assertListEqual(A , A ) # Testing the unknown token _UpperCAmelCase : List[str] = tokens + [rust_tokenizer.unk_token] _UpperCAmelCase : int = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A ) , A ) def __lowerCAmelCase ( self , *A , **A ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def __lowerCAmelCase ( self , A=1_5 ) -> List[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): _UpperCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(A , **A ) # Simple input _UpperCAmelCase : str = '''This is a simple input''' _UpperCAmelCase : Tuple = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCAmelCase : Optional[Any] = ('''This is a simple input''', '''This is a pair''') _UpperCAmelCase : str = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding='''max_length''' ) # Simple input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding='''max_length''' ) # Simple input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding='''max_length''' , ) # Pair input self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding='''max_length''' ) # Pair input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding='''max_length''' ) # Pair input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding='''max_length''' , ) def __lowerCAmelCase ( self ) -> List[str]: _UpperCAmelCase : Optional[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input _UpperCAmelCase : Optional[int] = '''This is a simple input''' _UpperCAmelCase : Dict = ['''This is a simple input looooooooong''', '''This is a simple input'''] _UpperCAmelCase : Union[str, Any] = ('''This is a simple input''', '''This is a pair''') _UpperCAmelCase : Optional[Any] = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] _UpperCAmelCase : List[str] = tokenizer.pad_token_id _UpperCAmelCase : Tuple = tokenizer(A , padding='''max_length''' , max_length=3_0 , return_tensors='''np''' ) _UpperCAmelCase : Optional[Any] = tokenizer(A , padding=A , truncate=A , return_tensors='''np''' ) _UpperCAmelCase : int = tokenizer(*A , padding='''max_length''' , max_length=6_0 , return_tensors='''np''' ) _UpperCAmelCase : List[str] = tokenizer(A , padding=A , truncate=A , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def __lowerCAmelCase ( self ) -> List[str]: _UpperCAmelCase : Optional[int] = '''$$$''' _UpperCAmelCase : Optional[int] = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A , add_bos_token=A ) _UpperCAmelCase : Tuple = '''This is a simple input''' _UpperCAmelCase : int = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCAmelCase : List[str] = tokenizer.bos_token_id _UpperCAmelCase : str = tokenizer(A ) _UpperCAmelCase : Optional[Any] = tokenizer(A ) self.assertEqual(out_s.input_ids[0] , A ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _UpperCAmelCase : Tuple = tokenizer.decode(out_s.input_ids ) _UpperCAmelCase : List[str] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def __lowerCAmelCase ( self ) -> Optional[Any]: _UpperCAmelCase : Optional[Any] = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) _UpperCAmelCase : Any = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' _UpperCAmelCase : Union[str, Any] = '''\nif len_a > len_b: result = a\nelse: result = b''' _UpperCAmelCase : Any = tokenizer.encode(A ) _UpperCAmelCase : Tuple = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] _UpperCAmelCase : List[str] = tokenizer.decode(A , truncate_before_pattern=A ) self.assertEqual(A , A ) def __lowerCAmelCase ( self ) -> Optional[Any]: pass
506
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { """google/pegasus-large""": """https://huggingface.co/google/pegasus-large/resolve/main/config.json""", # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowerCamelCase : Optional[int] ="pegasus" lowerCamelCase : str =["past_key_values"] lowerCamelCase : str ={"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : int , lowerCAmelCase : Union[str, Any]=5_02_65 , lowerCAmelCase : Union[str, Any]=10_24 , lowerCAmelCase : List[Any]=12 , lowerCAmelCase : int=40_96 , lowerCAmelCase : List[Any]=16 , lowerCAmelCase : Tuple=12 , lowerCAmelCase : Tuple=40_96 , lowerCAmelCase : str=16 , lowerCAmelCase : str=0.0 , lowerCAmelCase : List[Any]=0.0 , lowerCAmelCase : Union[str, Any]=True , lowerCAmelCase : Tuple=True , lowerCAmelCase : List[Any]="gelu" , lowerCAmelCase : Optional[int]=10_24 , lowerCAmelCase : Optional[Any]=0.1 , lowerCAmelCase : Dict=0.0 , lowerCAmelCase : Optional[int]=0.0 , lowerCAmelCase : Dict=0.02 , lowerCAmelCase : str=0 , lowerCAmelCase : Any=False , lowerCAmelCase : List[Any]=0 , lowerCAmelCase : List[str]=1 , lowerCAmelCase : Optional[Any]=1 , **lowerCAmelCase : Tuple , ) -> Tuple: """simple docstring""" __lowerCAmelCase : Optional[Any] = vocab_size __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Optional[Any] = d_model __lowerCAmelCase : Any = encoder_ffn_dim __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : str = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_ffn_dim __lowerCAmelCase : Optional[int] = decoder_layers __lowerCAmelCase : int = decoder_attention_heads __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Optional[int] = attention_dropout __lowerCAmelCase : Any = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : int = init_std __lowerCAmelCase : int = encoder_layerdrop __lowerCAmelCase : Tuple = decoder_layerdrop __lowerCAmelCase : str = use_cache __lowerCAmelCase : str = encoder_layers __lowerCAmelCase : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , is_encoder_decoder=lowerCAmelCase , decoder_start_token_id=lowerCAmelCase , forced_eos_token_id=lowerCAmelCase , **lowerCAmelCase , ) @property def SCREAMING_SNAKE_CASE ( self : Any ) -> int: """simple docstring""" return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: """simple docstring""" return self.d_model
218
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self : Union[str, Any] , lowerCAmelCase : List[str] , lowerCAmelCase : Tuple=7 , lowerCAmelCase : Optional[Any]=3 , lowerCAmelCase : str=30 , lowerCAmelCase : Optional[Any]=4_00 , lowerCAmelCase : int=True , lowerCAmelCase : Tuple=None , lowerCAmelCase : int=True , lowerCAmelCase : Optional[Any]=[0.5, 0.5, 0.5] , lowerCAmelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCAmelCase : List[str]=True , lowerCAmelCase : List[str]=1 / 2_55 , lowerCAmelCase : Any=True , ) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 13_33} __lowerCAmelCase : Any = parent __lowerCAmelCase : Union[str, Any] = batch_size __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = min_resolution __lowerCAmelCase : int = max_resolution __lowerCAmelCase : Any = do_resize __lowerCAmelCase : Tuple = size __lowerCAmelCase : str = do_normalize __lowerCAmelCase : Dict = image_mean __lowerCAmelCase : Optional[int] = image_std __lowerCAmelCase : List[str] = do_rescale __lowerCAmelCase : List[str] = rescale_factor __lowerCAmelCase : str = do_pad def SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[Any]: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase : Any , lowerCAmelCase : Dict=False ) -> str: """simple docstring""" if not batched: __lowerCAmelCase : Dict = image_inputs[0] if isinstance(lowerCAmelCase , Image.Image ): __lowerCAmelCase ,__lowerCAmelCase : str = image.size else: __lowerCAmelCase ,__lowerCAmelCase : Any = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase : List[str] = int(self.size["""shortest_edge"""] * h / w ) __lowerCAmelCase : List[Any] = self.size["""shortest_edge"""] elif w > h: __lowerCAmelCase : Union[str, Any] = self.size["""shortest_edge"""] __lowerCAmelCase : List[Any] = int(self.size["""shortest_edge"""] * w / h ) else: __lowerCAmelCase : Optional[Any] = self.size["""shortest_edge"""] __lowerCAmelCase : Any = self.size["""shortest_edge"""] else: __lowerCAmelCase : Optional[Any] = [] for image in image_inputs: __lowerCAmelCase ,__lowerCAmelCase : Optional[int] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase : Union[str, Any] = max(lowerCAmelCase , key=lambda lowerCAmelCase : item[0] )[0] __lowerCAmelCase : List[str] = max(lowerCAmelCase , key=lambda lowerCAmelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowerCamelCase : Dict =YolosImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: """simple docstring""" __lowerCAmelCase : Any = YolosImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase , """image_mean""" ) ) self.assertTrue(hasattr(lowerCAmelCase , """image_std""" ) ) self.assertTrue(hasattr(lowerCAmelCase , """do_normalize""" ) ) self.assertTrue(hasattr(lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(lowerCAmelCase , """size""" ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 13_33} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase ) __lowerCAmelCase : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCAmelCase ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Tuple: """simple docstring""" pass def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Dict: """simple docstring""" __lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , Image.Image ) # Test not batched input __lowerCAmelCase : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase ,__lowerCAmelCase : str = self.image_processor_tester.get_expected_values(lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase ,__lowerCAmelCase : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCAmelCase , batched=lowerCAmelCase ) __lowerCAmelCase : Optional[int] = image_processing(lowerCAmelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , numpify=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , np.ndarray ) # Test not batched input __lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase ,__lowerCAmelCase : str = self.image_processor_tester.get_expected_values(lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : Any = image_processing(lowerCAmelCase , return_tensors="""pt""" ).pixel_values __lowerCAmelCase ,__lowerCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(lowerCAmelCase , batched=lowerCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Dict: """simple docstring""" __lowerCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , torchify=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , torch.Tensor ) # Test not batched input __lowerCAmelCase : str = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase ,__lowerCAmelCase : Dict = self.image_processor_tester.get_expected_values(lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : Optional[int] = image_processing(lowerCAmelCase , return_tensors="""pt""" ).pixel_values __lowerCAmelCase ,__lowerCAmelCase : List[str] = self.image_processor_tester.get_expected_values(lowerCAmelCase , batched=lowerCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) __lowerCAmelCase : List[str] = self.image_processing_class(do_resize=lowerCAmelCase , do_normalize=lowerCAmelCase , do_rescale=lowerCAmelCase ) # create random PyTorch tensors __lowerCAmelCase : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , torchify=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors __lowerCAmelCase : Optional[Any] = image_processing_a.pad(lowerCAmelCase , return_tensors="""pt""" ) __lowerCAmelCase : Optional[int] = image_processing_a(lowerCAmelCase , return_tensors="""pt""" ) self.assertTrue( torch.allclose(encoded_images_with_method["""pixel_values"""] , encoded_images["""pixel_values"""] , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: __lowerCAmelCase : List[str] = json.loads(f.read() ) __lowerCAmelCase : Tuple = {"""image_id""": 3_97_69, """annotations""": target} # encode them __lowerCAmelCase : int = YolosImageProcessor.from_pretrained("""hustvl/yolos-small""" ) __lowerCAmelCase : List[Any] = image_processing(images=lowerCAmelCase , annotations=lowerCAmelCase , return_tensors="""pt""" ) # verify pixel values __lowerCAmelCase : Dict = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , lowerCAmelCase ) __lowerCAmelCase : Dict = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , lowerCAmelCase , atol=1e-4 ) ) # verify area __lowerCAmelCase : Union[str, Any] = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , lowerCAmelCase ) ) # verify boxes __lowerCAmelCase : Union[str, Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , lowerCAmelCase ) __lowerCAmelCase : Dict = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , lowerCAmelCase , atol=1e-3 ) ) # verify image_id __lowerCAmelCase : Optional[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , lowerCAmelCase ) ) # verify is_crowd __lowerCAmelCase : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , lowerCAmelCase ) ) # verify class_labels __lowerCAmelCase : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , lowerCAmelCase ) ) # verify orig_size __lowerCAmelCase : str = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , lowerCAmelCase ) ) # verify size __lowerCAmelCase : Optional[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , lowerCAmelCase ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: """simple docstring""" __lowerCAmelCase : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: __lowerCAmelCase : int = json.loads(f.read() ) __lowerCAmelCase : str = {"""file_name""": """000000039769.png""", """image_id""": 3_97_69, """segments_info""": target} __lowerCAmelCase : List[str] = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them __lowerCAmelCase : Any = YolosImageProcessor(format="""coco_panoptic""" ) __lowerCAmelCase : str = image_processing(images=lowerCAmelCase , annotations=lowerCAmelCase , masks_path=lowerCAmelCase , return_tensors="""pt""" ) # verify pixel values __lowerCAmelCase : Tuple = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , lowerCAmelCase ) __lowerCAmelCase : List[str] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , lowerCAmelCase , atol=1e-4 ) ) # verify area __lowerCAmelCase : Union[str, Any] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , lowerCAmelCase ) ) # verify boxes __lowerCAmelCase : Optional[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , lowerCAmelCase ) __lowerCAmelCase : int = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , lowerCAmelCase , atol=1e-3 ) ) # verify image_id __lowerCAmelCase : Tuple = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , lowerCAmelCase ) ) # verify is_crowd __lowerCAmelCase : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , lowerCAmelCase ) ) # verify class_labels __lowerCAmelCase : Optional[Any] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , lowerCAmelCase ) ) # verify masks __lowerCAmelCase : List[Any] = 82_28_73 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , lowerCAmelCase ) # verify orig_size __lowerCAmelCase : Dict = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , lowerCAmelCase ) ) # verify size __lowerCAmelCase : Tuple = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , lowerCAmelCase ) )
218
1
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case_ = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class snake_case_ ( _A , unittest.TestCase): lowerCamelCase :Optional[Any] = XGLMTokenizer lowerCamelCase :Optional[Any] = XGLMTokenizerFast lowerCamelCase :List[Any] = True lowerCamelCase :Union[str, Any] = True def __lowercase ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase : Optional[Any] =XGLMTokenizer(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self ) -> Any: lowerCamelCase : Optional[int] ='''<pad>''' lowerCamelCase : int =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def __lowercase ( self ) -> List[str]: lowerCamelCase : int =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(len(__lowercase ) , 1_0_0_8 ) def __lowercase ( self ) -> int: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_8 ) def __lowercase ( self ) -> Optional[Any]: lowerCamelCase : Optional[Any] =XGLMTokenizer(__lowercase , keep_accents=__lowercase ) lowerCamelCase : List[str] =tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCamelCase : int =tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCamelCase : List[Any] =tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] , ) lowerCamelCase : Dict =tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def __lowercase ( self ) -> Optional[Any]: return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def __lowercase ( self ) -> Union[str, Any]: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__lowercase , f.name ) lowerCamelCase : List[Any] =XGLMTokenizer(f.name , keep_accents=__lowercase ) lowerCamelCase : Optional[int] =pickle.dumps(__lowercase ) pickle.loads(__lowercase ) def __lowercase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return lowerCamelCase : int =self.get_tokenizer() lowerCamelCase : List[Any] =self.get_rust_tokenizer() lowerCamelCase : Dict ='''I was born in 92000, and this is falsé.''' lowerCamelCase : Optional[int] =tokenizer.tokenize(__lowercase ) lowerCamelCase : Optional[int] =rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) lowerCamelCase : List[Any] =tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) lowerCamelCase : Any =rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) lowerCamelCase : Union[str, Any] =self.get_rust_tokenizer() lowerCamelCase : Any =tokenizer.encode(__lowercase ) lowerCamelCase : Tuple =rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def __lowercase ( self ) -> Tuple: lowerCamelCase : Union[str, Any] ='''Hello World!''' lowerCamelCase : Dict =[2, 3_1_2_2_7, 4_4_4_7, 3_5] self.assertListEqual(__lowercase , self.big_tokenizer.encode(__lowercase ) ) @slow def __lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Any =( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off lowerCamelCase : List[str] =[2, 1_0_1_8, 6_7, 1_1, 1_9_8_8, 2_6_1_7, 5_6_3_1, 2_7_8, 1_1, 3_4_0_7, 4_8, 7_1_6_3_0, 2_8_0_8_5, 4, 3_2_3_4, 1_5_7, 1_3, 6, 5, 6, 4, 3_5_2_6, 7_6_8, 1_5, 6_5_9, 5_7, 2_9_8, 3_9_8_3, 8_6_4, 1_2_9, 2_1, 6, 5, 1_3_6_7_5, 3_7_7, 6_5_2, 7_5_8_0, 1_0_3_4_1, 1_5_5, 2_8_1_7, 4_2_2, 1_6_6_6, 7, 1_6_7_4, 5_3, 1_1_3, 2_0_2_2_7_7, 1_7_8_9_2, 3_3, 6_0, 8_7, 4, 3_2_3_4, 1_5_7, 6_1, 2_6_6_7, 5_2_3_7_6, 1_9, 8_8, 2_3, 7_3_5] # fmt: on self.assertListEqual(__lowercase , self.big_tokenizer.encode(__lowercase ) ) @slow def __lowercase ( self ) -> Tuple: # fmt: off lowerCamelCase : List[str] ={ '''input_ids''': [[2, 1_0_8_8_2_5, 1_1_6_3, 1_5, 8_8_0_1_0, 4_7_3, 1_5_8_9_8, 1_5_7, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 2_3_8_0_2_1, 1_1_6_3, 5_3, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 5_3_2_8_3, 1_8_2_3_9_6, 8, 1_8_5_6_6, 1_6, 3_6_7_3_3, 4_1_0_1, 8, 2_3_0, 2_4_4_0_1_7, 1_2_2_5_5_3, 7, 1_5, 1_3_2_5_9_7, 4, 2_9_3, 1_2_5_1_1, 7_6_1_0, 4, 3_4_1_4, 1_3_2_5_9_7, 9, 4, 3_2_3_6_1, 3_6_2, 4, 7_3_4, 2_8_5_1_2, 3_2_5_6_9, 1_8, 4, 3_2_3_6_1, 2_6_0_9_6, 1_4_9_8_2, 7_3, 1_8_7_1_5, 2_1_4_3_3, 2_3_5_2_6_1, 1_5, 4_9_2, 1_2_4_2_7, 1_6, 5_3, 1_8_7_1_5, 2_1_4_3_3, 6_5_4_5_4, 1_5, 2_3_6_5_9, 5_6_3, 1_6, 2_7_8, 5_9_7, 2_8_4_3, 5_9_5, 7_9_3_1, 1_8_2_3_9_6, 6_4_1_8_6, 2_2, 8_8_6, 5_9_5, 1_3_2_9_8_1, 5_3, 2_5_5_4_0, 3_4_4_9, 4_3_9_8_2, 3_9_9_0_1, 5_9_5_1, 8_7_8, 3_3_0, 4, 2_7_6_9_4, 8_0_2_6_9, 3_1_2, 5_3, 6_5_1_7, 1_1_7_8_0, 6_1_1, 2_0_4_0_8, 5], [2, 6, 1_3_2_5_9_7, 6_7, 4_2_8_9_7, 3_3, 5_9_2, 8, 1_6_3_7_2_9, 2_5_5_4_0, 3_6_1, 1_3_6_9_9_7, 1_0_9_5_1_4, 1_7_3_2_3_0, 7, 5_0_1, 6_0, 1_0_2_9_1_3, 1_9_6, 5_6_3_1, 2_3_5, 6_3_2_4_3, 4_7_3, 6, 2_3_1_7_5_7, 7_4, 5_2_7_7, 7_9_0_5, 5_3, 3_0_9_5, 3_7_3_1_7, 2_2, 4_5_4, 1_8_3_8_7_4, 5], [2, 2_6_8, 3_1_2_9_8, 4_6_5_3_0, 6, 1_3_2_9_3_5, 4_3_8_3_1, 7, 5_9_7, 3_2, 2_4, 3_6_8_8, 9_8_6_5, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name='''facebook/xglm-564M''' , padding=__lowercase , )
262
def A__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 1_0 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_3_1_7_0_4_4_0_6_4_6_7_9_8_8_7_3_8_5_9_6_1_9_8_1 and not allow_probable: raise ValueError( '''Warning: upper bound of deterministic test is exceeded. ''' '''Pass allow_probable=True to allow probabilistic test. ''' '''A return value of True indicates a probable prime.''' ) # array bounds provided by analysis lowerCamelCase : Optional[Any] =[ 2_0_4_7, 1_3_7_3_6_5_3, 2_5_3_2_6_0_0_1, 3_2_1_5_0_3_1_7_5_1, 2_1_5_2_3_0_2_8_9_8_7_4_7, 3_4_7_4_7_4_9_6_6_0_3_8_3, 3_4_1_5_5_0_0_7_1_7_2_8_3_2_1, 1, 3_8_2_5_1_2_3_0_5_6_5_4_6_4_1_3_0_5_1, 1, 1, 3_1_8_6_6_5_8_5_7_8_3_4_0_3_1_1_5_1_1_6_7_4_6_1, 3_3_1_7_0_4_4_0_6_4_6_7_9_8_8_7_3_8_5_9_6_1_9_8_1, ] lowerCamelCase : Dict =[2, 3, 5, 7, 1_1, 1_3, 1_7, 1_9, 2_3, 2_9, 3_1, 3_7, 4_1] for idx, _p in enumerate(SCREAMING_SNAKE_CASE_ , 1 ): if n < _p: # then we have our last prime to check lowerCamelCase : Any =primes[:idx] break lowerCamelCase , lowerCamelCase : Any =n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: lowerCamelCase : Union[str, Any] =False for r in range(SCREAMING_SNAKE_CASE_ ): lowerCamelCase : List[str] =pow(SCREAMING_SNAKE_CASE_ , d * 2**r , SCREAMING_SNAKE_CASE_ ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): lowerCamelCase : List[str] =True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def A__ ( ) -> None: assert not miller_rabin(5_6_1 ) assert miller_rabin(5_6_3 ) # 2047 assert not miller_rabin(8_3_8_2_0_1 ) assert miller_rabin(8_3_8_2_0_7 ) # 1_373_653 assert not miller_rabin(1_7_3_1_6_0_0_1 ) assert miller_rabin(1_7_3_1_6_0_1_7 ) # 25_326_001 assert not miller_rabin(3_0_7_8_3_8_6_6_4_1 ) assert miller_rabin(3_0_7_8_3_8_6_6_5_3 ) # 3_215_031_751 assert not miller_rabin(1_7_1_3_0_4_5_5_7_4_8_0_1 ) assert miller_rabin(1_7_1_3_0_4_5_5_7_4_8_1_9 ) # 2_152_302_898_747 assert not miller_rabin(2_7_7_9_7_9_9_7_2_8_3_0_7 ) assert miller_rabin(2_7_7_9_7_9_9_7_2_8_3_2_7 ) # 3_474_749_660_383 assert not miller_rabin(1_1_3_8_5_0_0_2_3_9_0_9_4_4_1 ) assert miller_rabin(1_1_3_8_5_0_0_2_3_9_0_9_5_2_7 ) # 341_550_071_728_321 assert not miller_rabin(1_2_7_5_0_4_1_0_1_8_8_4_8_8_0_4_3_5_1 ) assert miller_rabin(1_2_7_5_0_4_1_0_1_8_8_4_8_8_0_4_3_9_1 ) # 3_825_123_056_546_413_051 assert not miller_rabin(7_9_6_6_6_4_6_4_4_5_8_5_0_7_7_8_7_7_9_1_8_6_7 ) assert miller_rabin(7_9_6_6_6_4_6_4_4_5_8_5_0_7_7_8_7_7_9_1_9_5_1 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(5_5_2_8_4_0_6_7_7_4_4_6_6_4_7_8_9_7_6_6_0_3_3_3 ) assert miller_rabin(5_5_2_8_4_0_6_7_7_4_4_6_6_4_7_8_9_7_6_6_0_3_5_9 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
262
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCamelCase_ ) class lowerCAmelCase_ ( lowerCamelCase_ ): __a : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True} ) __a : ClassVar[Features] = Features({"text": Value("string" )} ) __a : ClassVar[Features] = Features({} ) __a : str = "text" @property def snake_case ( self ): return {self.text_column: "text"}
105
'''simple docstring''' import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def __UpperCAmelCase ( SCREAMING_SNAKE_CASE__: Optional[Any], SCREAMING_SNAKE_CASE__: Tuple, SCREAMING_SNAKE_CASE__: Dict ) -> List[Any]: """simple docstring""" # Initialise PyTorch model __a = RemBertConfig.from_json_file(SCREAMING_SNAKE_CASE__ ) print('Building PyTorch model from configuration: {}'.format(str(SCREAMING_SNAKE_CASE__ ) ) ) __a = RemBertModel(SCREAMING_SNAKE_CASE__ ) # Load weights from tf checkpoint load_tf_weights_in_rembert(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) # Save pytorch-model print('Save PyTorch model to {}'.format(SCREAMING_SNAKE_CASE__ ) ) torch.save(model.state_dict(), SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __UpperCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--rembert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained RemBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __UpperCamelCase : List[str] = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
448
0
"""simple docstring""" import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def a__ ( __lowercase ) -> int: _A = [] embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", f"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", f"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", f"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", f"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def a__ ( __lowercase , __lowercase ) -> str: _A = [] attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", f"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", f"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", f"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", f"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", f"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", f"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", f"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", f"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def a__ ( __lowercase ) -> Dict: _A = [] token.append((f"""cvt.encoder.stages.{idx}.cls_token""", "stage2.cls_token") ) return token def a__ ( ) -> Dict: _A = [] head.append(("layernorm.weight", "norm.weight") ) head.append(("layernorm.bias", "norm.bias") ) head.append(("classifier.weight", "head.weight") ) head.append(("classifier.bias", "head.bias") ) return head def a__ ( __lowercase , __lowercase , __lowercase , __lowercase ) -> Optional[Any]: _A = "imagenet-1k-id2label.json" _A = 1000 _A = "huggingface/label-files" _A = num_labels _A = json.load(open(cached_download(hf_hub_url(__lowercase , __lowercase , repo_type="dataset" ) ) , "r" ) ) _A = {int(__lowercase ): v for k, v in idalabel.items()} _A = idalabel _A = {v: k for k, v in idalabel.items()} _A = _A = CvtConfig(num_labels=__lowercase , idalabel=__lowercase , labelaid=__lowercase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit("/" , 1 )[-1][4:6] == "13": _A = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit("/" , 1 )[-1][4:6] == "21": _A = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: _A = [2, 2, 20] _A = [3, 12, 16] _A = [192, 768, 1024] _A = CvtForImageClassification(__lowercase ) _A = AutoImageProcessor.from_pretrained("facebook/convnext-base-224-22k-1k" ) _A = image_size _A = torch.load(__lowercase , map_location=torch.device("cpu" ) ) _A = OrderedDict() _A = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: _A = list_of_state_dict + cls_token(__lowercase ) _A = list_of_state_dict + embeddings(__lowercase ) for cnt in range(config.depth[idx] ): _A = list_of_state_dict + attention(__lowercase , __lowercase ) _A = list_of_state_dict + final() for gg in list_of_state_dict: print(__lowercase ) for i in range(len(__lowercase ) ): _A = original_weights[list_of_state_dict[i][1]] model.load_state_dict(__lowercase ) model.save_pretrained(__lowercase ) image_processor.save_pretrained(__lowercase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( "--cvt_model", default="cvt-w24", type=str, help="Name of the cvt model you'd like to convert.", ) parser.add_argument( "--image_size", default=3_84, type=int, help="Input Image Size", ) parser.add_argument( "--cvt_file_name", default=r"cvtmodels\CvT-w24-384x384-IN-22k.pth", type=str, help="Input Image Size", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) a_ = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
707
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging a_ = logging.get_logger(__name__) class snake_case ( _UpperCamelCase): __UpperCamelCase = ['input_features'] def __init__( self : int , a__ : Optional[Any]=80 , a__ : Optional[int]=1_60_00 , a__ : int=1_60 , a__ : Union[str, Any]=30 , a__ : Tuple=4_00 , a__ : List[Any]=0.0 , a__ : Optional[Any]=False , **a__ : List[Any] , ) -> str: '''simple docstring''' super().__init__( feature_size=a__ , sampling_rate=a__ , padding_value=a__ , return_attention_mask=a__ , **a__ , ) _A = n_fft _A = hop_length _A = chunk_length _A = chunk_length * sampling_rate _A = self.n_samples // hop_length _A = sampling_rate _A = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a__ , min_frequency=0.0 , max_frequency=8_0_0_0.0 , sampling_rate=a__ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : int , a__ : np.array ) -> np.ndarray: '''simple docstring''' _A = spectrogram( a__ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="log10" , ) _A = log_spec[:, :-1] _A = np.maximum(a__ , log_spec.max() - 8.0 ) _A = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def a_ ( a__ : List[np.ndarray] , a__ : List[np.ndarray] , a__ : float = 0.0 ) -> List[np.ndarray]: '''simple docstring''' if attention_mask is not None: _A = np.array(a__ , np.intaa ) _A = [] for vector, length in zip(a__ , attention_mask.sum(-1 ) ): _A = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _A = padding_value normed_input_values.append(a__ ) else: _A = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self : Optional[int] , a__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a__ : bool = True , a__ : Optional[int] = None , a__ : Optional[Union[str, TensorType]] = None , a__ : Optional[bool] = None , a__ : Optional[str] = "max_length" , a__ : Optional[int] = None , a__ : Optional[int] = None , a__ : Optional[bool] = None , **a__ : Dict , ) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"""The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a""" F""" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input""" F""" was sampled with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) _A = isinstance(a__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) _A = is_batched_numpy or ( isinstance(a__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _A = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a__ , np.ndarray ): _A = np.asarray(a__ , dtype=np.floataa ) elif isinstance(a__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _A = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _A = [np.asarray([raw_speech] ).T] _A = BatchFeature({"input_features": raw_speech} ) # convert into correct format for padding _A = self.pad( a__ , padding=a__ , max_length=max_length if max_length else self.n_samples , truncation=a__ , pad_to_multiple_of=a__ , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: _A = self.zero_mean_unit_var_norm( padded_inputs["input_features"] , attention_mask=padded_inputs["attention_mask"] , padding_value=self.padding_value , ) _A = np.stack(padded_inputs["input_features"] , axis=0 ) # make sure list is in array format _A = padded_inputs.get("input_features" ).transpose(2 , 0 , 1 ) _A = [self._np_extract_fbank_features(a__ ) for waveform in input_features[0]] if isinstance(input_features[0] , a__ ): _A = [np.asarray(a__ , dtype=np.floataa ) for feature in input_features] else: _A = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _A = padded_inputs["attention_mask"][:, :: self.hop_length] if return_tensors is not None: _A = padded_inputs.convert_to_tensors(a__ ) return padded_inputs def a_ ( self : Dict ) -> Dict[str, Any]: '''simple docstring''' _A = copy.deepcopy(self.__dict__ ) _A = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
621
0
'''simple docstring''' import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCamelCase__ ( A : Tuple , A : str , A : int , A : Optional[int] , A : Optional[Any]=True , A : Dict="pt" ): '''simple docstring''' UpperCAmelCase = {'''add_prefix_space''': True} if isinstance(A , A ) and not line.startswith(''' ''' ) else {} UpperCAmelCase = padding_side return tokenizer( [line] , max_length=A , padding='''max_length''' if pad_to_max_length else None , truncation=A , return_tensors=A , add_special_tokens=A , **A , ) def lowerCamelCase__ ( A : List[Any] , A : Optional[int] , A : Tuple=None , ): '''simple docstring''' UpperCAmelCase = input_ids.ne(A ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class UpperCamelCase__( lowerCAmelCase ): def __init__( self : Tuple , lowerCAmelCase : Optional[int] , lowerCAmelCase : int , lowerCAmelCase : Optional[int] , lowerCAmelCase : int , lowerCAmelCase : Tuple="train" , lowerCAmelCase : List[Any]=None , lowerCAmelCase : Tuple=None , lowerCAmelCase : List[Any]=None , lowerCAmelCase : Union[str, Any]="" , )-> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase = Path(lowerCAmelCase ).joinpath(type_path + '''.source''' ) UpperCAmelCase = Path(lowerCAmelCase ).joinpath(type_path + '''.target''' ) UpperCAmelCase = self.get_char_lens(self.src_file ) UpperCAmelCase = max_source_length UpperCAmelCase = max_target_length assert min(self.src_lens ) > 0, F"""found empty line in {self.src_file}""" UpperCAmelCase = tokenizer UpperCAmelCase = prefix if n_obs is not None: UpperCAmelCase = self.src_lens[:n_obs] UpperCAmelCase = src_lang UpperCAmelCase = tgt_lang def __len__( self : Tuple )-> Optional[Any]: """simple docstring""" return len(self.src_lens ) def __getitem__( self : List[Any] , lowerCAmelCase : List[str] )-> Dict[str, torch.Tensor]: """simple docstring""" UpperCAmelCase = index + 1 # linecache starts at 1 UpperCAmelCase = self.prefix + linecache.getline(str(self.src_file ) , lowerCAmelCase ).rstrip('''\n''' ) UpperCAmelCase = linecache.getline(str(self.tgt_file ) , lowerCAmelCase ).rstrip('''\n''' ) assert source_line, F"""empty source line for index {index}""" assert tgt_line, F"""empty tgt line for index {index}""" # Need to add eos token manually for T5 if isinstance(self.tokenizer , lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , lowerCAmelCase ) else self.tokenizer ) UpperCAmelCase = self.tokenizer.generator if isinstance(self.tokenizer , lowerCAmelCase ) else self.tokenizer UpperCAmelCase = encode_line(lowerCAmelCase , lowerCAmelCase , self.max_source_length , '''right''' ) UpperCAmelCase = encode_line(lowerCAmelCase , lowerCAmelCase , self.max_target_length , '''right''' ) UpperCAmelCase = source_inputs['''input_ids'''].squeeze() UpperCAmelCase = target_inputs['''input_ids'''].squeeze() UpperCAmelCase = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def a__( lowerCAmelCase : str )-> int: """simple docstring""" return [len(lowerCAmelCase ) for x in Path(lowerCAmelCase ).open().readlines()] def a__( self : List[str] , lowerCAmelCase : Dict )-> Dict[str, torch.Tensor]: """simple docstring""" UpperCAmelCase = torch.stack([x['''input_ids'''] for x in batch] ) UpperCAmelCase = torch.stack([x['''attention_mask'''] for x in batch] ) UpperCAmelCase = torch.stack([x['''decoder_input_ids'''] for x in batch] ) UpperCAmelCase = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , lowerCAmelCase ) else self.tokenizer.pad_token_id ) UpperCAmelCase = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , lowerCAmelCase ) else self.tokenizer.pad_token_id ) UpperCAmelCase = trim_batch(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase = trim_batch(lowerCAmelCase , lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch _lowercase : Tuple = getLogger(__name__) def lowerCamelCase__ ( A : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(A ) ) def lowerCamelCase__ ( A : str ): '''simple docstring''' UpperCAmelCase = get_git_info() save_json(A , os.path.join(A , '''git_log.json''' ) ) def lowerCamelCase__ ( A : List[str] , A : Dict , A : Optional[int]=4 , **A : Union[str, Any] ): '''simple docstring''' with open(A , '''w''' ) as f: json.dump(A , A , indent=A , **A ) def lowerCamelCase__ ( A : Dict ): '''simple docstring''' with open(A ) as f: return json.load(A ) def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase = git.Repo(search_parent_directories=A ) UpperCAmelCase = { '''repo_id''': str(A ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def lowerCamelCase__ ( A : Callable , A : Iterable ): '''simple docstring''' return list(map(A , A ) ) def lowerCamelCase__ ( A : List[str] , A : Union[str, Any] ): '''simple docstring''' with open(A , '''wb''' ) as f: return pickle.dump(A , A ) def lowerCamelCase__ ( A : Optional[int] ): '''simple docstring''' def remove_articles(A : int ): return re.sub(R'''\b(a|an|the)\b''' , ''' ''' , A ) def white_space_fix(A : List[Any] ): return " ".join(text.split() ) def remove_punc(A : Dict ): UpperCAmelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(A : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(A ) ) ) ) def lowerCamelCase__ ( A : Tuple , A : List[Any] ): '''simple docstring''' UpperCAmelCase = normalize_answer(A ).split() UpperCAmelCase = normalize_answer(A ).split() UpperCAmelCase = Counter(A ) & Counter(A ) UpperCAmelCase = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase = 1.0 * num_same / len(A ) UpperCAmelCase = 1.0 * num_same / len(A ) UpperCAmelCase = (2 * precision * recall) / (precision + recall) return fa def lowerCamelCase__ ( A : Union[str, Any] , A : Tuple ): '''simple docstring''' return normalize_answer(A ) == normalize_answer(A ) def lowerCamelCase__ ( A : List[str] , A : List[str] ): '''simple docstring''' assert len(A ) == len(A ) UpperCAmelCase = 0 for hypo, pred in zip(A , A ): em += exact_match_score(A , A ) if len(A ) > 0: em /= len(A ) return {"em": em} def lowerCamelCase__ ( A : Union[str, Any] ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def lowerCamelCase__ ( A : str , A : Any , A : Tuple ): '''simple docstring''' UpperCAmelCase = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase = '''dropout_rate''' for p in extra_params: if getattr(A , A , A ): if not hasattr(A , A ) and not hasattr(A , equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(A ) ) delattr(A , A ) continue UpperCAmelCase = p if hasattr(A , A ) else equivalent_param[p] setattr(A , A , getattr(A , A ) ) delattr(A , A ) return hparams, config
210
'''simple docstring''' import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _lowercase : Tuple = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. _lowercase : str = importlib.util.spec_from_file_location( """transformers""", os.path.join(PATH_TO_TRANSFORMERS, """__init__.py"""), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) _lowercase : Optional[Any] = spec.loader.load_module() _lowercase : List[Any] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` _lowercase : List[Any] = re.compile("""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") _lowercase : List[str] = { """CLIPConfigMixin""", """DecisionTransformerConfigMixin""", """EncoderDecoderConfigMixin""", """RagConfigMixin""", """SpeechEncoderDecoderConfigMixin""", """VisionEncoderDecoderConfigMixin""", """VisionTextDualEncoderConfigMixin""", } def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase = [] for config_class in list(CONFIG_MAPPING.values() ): UpperCAmelCase = False # source code of `config_class` UpperCAmelCase = inspect.getsource(A ) UpperCAmelCase = _re_checkpoint.findall(A ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` UpperCAmelCase , UpperCAmelCase = checkpoint # verify the checkpoint name corresponds to the checkpoint link UpperCAmelCase = f"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: UpperCAmelCase = True break UpperCAmelCase = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(A ) if len(A ) > 0: UpperCAmelCase = '''\n'''.join(sorted(A ) ) raise ValueError(f"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
210
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process __a : Optional[Any] = logging.getLogger(__name__) def __magic_name__ ( lowercase_ , lowercase_ ) -> List[str]: '''simple docstring''' return (preds == labels).mean() @dataclass class __UpperCAmelCase : """simple docstring""" lowercase = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase = field( default=snake_case__ , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase = field( default=snake_case__ , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase = field( default=snake_case__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class __UpperCAmelCase : """simple docstring""" lowercase = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys() )} ) lowercase = field(metadata={"""help""": """Should contain the data files for the task."""} ) lowercase = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowercase = field( default=snake_case__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def __magic_name__ ( ) -> str: '''simple docstring''' UpperCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCamelCase , UpperCamelCase , UpperCamelCase = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( "Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("Training/evaluation parameters %s" , lowercase_ ) # Set seed set_seed(training_args.seed ) try: UpperCamelCase = processors[data_args.task_name]() UpperCamelCase = processor.get_labels() UpperCamelCase = len(lowercase_ ) except KeyError: raise ValueError("Task not found: %s" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCamelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowercase_ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) UpperCamelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) UpperCamelCase = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , ) # Get datasets UpperCamelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowercase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) UpperCamelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowercase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(lowercase_ ) -> Dict: UpperCamelCase = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(lowercase_ , p.label_ids )} # Data collator UpperCamelCase = DataCollatorWithPadding(lowercase_ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer UpperCamelCase = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=lowercase_ , eval_dataset=lowercase_ , compute_metrics=lowercase_ , data_collator=lowercase_ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) UpperCamelCase = trainer.evaluate() UpperCamelCase = os.path.join(training_args.output_dir , "eval_results.txt" ) if trainer.is_world_master(): with open(lowercase_ , "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(" %s = %s" , lowercase_ , lowercase_ ) writer.write("%s = %s\n" % (key, value) ) results.update(lowercase_ ) return results def __magic_name__ ( lowercase_ ) -> Tuple: '''simple docstring''' main() if __name__ == "__main__": main()
414
import os import jsonlines import numpy as np from tqdm import tqdm __a : int = 2_0_4_8 __a : Optional[int] = 4_0_9_6 __a : Optional[int] = 4_2 __a : Optional[Any] = os.environ.pop("""PROCESS_TRAIN""", """false""") __a : Dict = {"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def __magic_name__ ( lowercase_ ) -> List[Any]: '''simple docstring''' def choose_first(lowercase_ , lowercase_=False ): assert isinstance(lowercase_ , lowercase_ ) if len(lowercase_ ) == 1: UpperCamelCase = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: UpperCamelCase = {k: [a[k]] for k in a} if len(a["start_token"] ) > 0: break return a UpperCamelCase = {"id": example["id"]} UpperCamelCase = example["annotations"] UpperCamelCase = annotation["yes_no_answer"] if 0 in yes_no_answer or 1 in yes_no_answer: UpperCamelCase = ["yes"] if 1 in yes_no_answer else ["no"] UpperCamelCase = UpperCamelCase = [] UpperCamelCase = UpperCamelCase = [] UpperCamelCase = ["<cls>"] else: UpperCamelCase = ["short"] UpperCamelCase = choose_first(annotation["short_answers"] ) if len(out["start_token"] ) == 0: # answer will be long if short is not available UpperCamelCase = ["long"] UpperCamelCase = choose_first(annotation["long_answer"] , is_long_answer=lowercase_ ) UpperCamelCase = [] answer.update(lowercase_ ) # disregard some samples if len(answer["start_token"] ) > 1 or answer["start_token"] == answer["end_token"]: UpperCamelCase = True else: UpperCamelCase = False UpperCamelCase = ["start_token", "end_token", "start_byte", "end_byte", "text"] if not all(isinstance(answer[k] , lowercase_ ) for k in cols ): raise ValueError("Issue in ID" , example["id"] ) return answer def __magic_name__ ( lowercase_ , lowercase_=False ) -> Optional[int]: '''simple docstring''' UpperCamelCase = _get_single_answer(lowercase_ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element UpperCamelCase = example["document"]["tokens"] UpperCamelCase = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) return { "context": " ".join(lowercase_ ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples UpperCamelCase = ["start_token", "end_token"] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 UpperCamelCase = example["document"]["tokens"] UpperCamelCase = answer["start_token"] UpperCamelCase = answer["end_token"] UpperCamelCase = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 UpperCamelCase = " ".join(context[start_token:end_token] ) # checking above code if assertion: UpperCamelCase = doc["is_html"][answer["start_token"] : answer["end_token"]] UpperCamelCase = doc["token"][answer["start_token"] : answer["end_token"]] UpperCamelCase = " ".join([old[i] for i in range(len(lowercase_ ) ) if not is_html[i]] ) if new != old: print("ID:" , example["id"] ) print("New:" , lowercase_ , end="\n" ) print("Old:" , lowercase_ , end="\n\n" ) return { "context": " ".join(lowercase_ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def __magic_name__ ( lowercase_ , lowercase_ , lowercase_=2048 , lowercase_=4096 , lowercase_=True ) -> int: '''simple docstring''' UpperCamelCase = get_context_and_ans(lowercase_ , assertion=lowercase_ ) UpperCamelCase = out["answer"] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } UpperCamelCase = tokenizer(example["question"]["text"] , out["context"] ).input_ids UpperCamelCase = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element UpperCamelCase = [] UpperCamelCase = [] UpperCamelCase = input_ids[:q_len] UpperCamelCase = range(lowercase_ , len(lowercase_ ) , max_length - doc_stride ) for i in doc_start_indices: UpperCamelCase = i + max_length - q_len UpperCamelCase = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer["category"][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(lowercase_ ), "end_token": [-100] * len(lowercase_ ), "category": category, }, } UpperCamelCase = out["context"].split() UpperCamelCase = splitted_context[answer["end_token"]] UpperCamelCase = len( tokenizer( " ".join(splitted_context[: answer["start_token"]] ) , add_special_tokens=lowercase_ , ).input_ids ) UpperCamelCase = len( tokenizer(" ".join(splitted_context[: answer["end_token"]] ) , add_special_tokens=lowercase_ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token UpperCamelCase = len(tokenizer(lowercase_ , add_special_tokens=lowercase_ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 UpperCamelCase = input_ids[answer["start_token"] : answer["end_token"] + 1] # right & left are inclusive UpperCamelCase = answer["start_token"] UpperCamelCase = answer["end_token"] if assertion: UpperCamelCase = tokenizer.decode(lowercase_ ) if answer["span"] != new: print("ISSUE IN TOKENIZATION" ) print("OLD:" , answer["span"] ) print("NEW:" , lowercase_ , end="\n\n" ) if len(lowercase_ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } UpperCamelCase = input_ids[:q_len] UpperCamelCase = range(lowercase_ , len(lowercase_ ) , max_length - doc_stride ) UpperCamelCase = [] UpperCamelCase = [] UpperCamelCase = [] UpperCamelCase = [] # null, yes, no, long, short for i in doc_start_indices: UpperCamelCase = i + max_length - q_len UpperCamelCase = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: UpperCamelCase = start_token - i + q_len UpperCamelCase = end_token - i + q_len answers_category.append(answer["category"][0] ) # ["short"] -> "short" else: UpperCamelCase = -100 UpperCamelCase = -100 answers_category.append("null" ) UpperCamelCase = inputs[-1][start_token : end_token + 1] answers_start_token.append(lowercase_ ) answers_end_token.append(lowercase_ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print("ISSUE in strided for ID:" , example["id"] ) print("New:" , tokenizer.decode(lowercase_ ) ) print("Old:" , tokenizer.decode(lowercase_ ) , end="\n\n" ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def __magic_name__ ( lowercase_ , lowercase_ , lowercase_=2048 , lowercase_=4096 , lowercase_=False ) -> Optional[Any]: '''simple docstring''' UpperCamelCase = get_strided_contexts_and_ans( lowercase_ , lowercase_ , doc_stride=lowercase_ , max_length=lowercase_ , assertion=lowercase_ , ) return example def __magic_name__ ( lowercase_ , lowercase_ ) -> Any: '''simple docstring''' with jsonlines.open(lowercase_ , "a" ) as writer: for example in tqdm(lowercase_ , total=len(lowercase_ ) , desc="Saving samples ... " ): UpperCamelCase = example["labels"] for ids, start, end, cat in zip( example["input_ids"] , labels["start_token"] , labels["end_token"] , labels["category"] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { "input_ids": ids, "start_token": start, "end_token": end, "category": CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer __a : Dict = load_dataset("""natural_questions""") __a : int = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") __a : List[Any] = data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] __a : Tuple = { """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } __a : int = data.map(prepare_inputs, fn_kwargs=fn_kwargs) __a : Optional[int] = data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) __a : List[Any] = """nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
414
1
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __magic_name__ (unittest.TestCase ): '''simple docstring''' __lowercase : Tuple = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:int , _a:str , _a:Optional[Any] ): snake_case__ = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) snake_case__ = VideoClassificationPipeline(model=_a , image_processor=_a , top_k=2 ) snake_case__ = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:Optional[Any] , _a:Optional[Any] ): for example in examples: snake_case__ = video_classifier(_a ) self.assertEqual( _a , [ {'''score''': ANY(_a ), '''label''': ANY(_a )}, {'''score''': ANY(_a ), '''label''': ANY(_a )}, ] , ) @require_torch def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' snake_case__ = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) snake_case__ = pipeline( '''video-classification''' , model=_a , feature_extractor=_a , frame_sampling_rate=4 ) snake_case__ = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) snake_case__ = video_classifier(_a , top_k=2 ) self.assertEqual( nested_simplify(_a , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) snake_case__ = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(_a , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass
33
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) A__ : int ={ '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : str =[ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Union[str, Any] =[ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : List[str] =[ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys A__ : Any =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
207
0
from math import log from scipy.constants import Boltzmann, physical_constants __UpperCAmelCase = 300 # TEMPERATURE (unit = K) def _lowerCamelCase ( A_ : float , A_ : float , A_ : float , ) -> float: '''simple docstring''' if donor_conc <= 0: raise ValueError("Donor concentration should be positive" ) elif acceptor_conc <= 0: raise ValueError("Acceptor concentration should be positive" ) elif intrinsic_conc <= 0: raise ValueError("Intrinsic concentration should be positive" ) elif donor_conc <= intrinsic_conc: raise ValueError( "Donor concentration should be greater than intrinsic concentration" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( "Acceptor concentration should be greater than intrinsic concentration" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
582
from math import factorial __UpperCAmelCase = {str(digit): factorial(digit) for digit in range(10)} def _lowerCamelCase ( A_ : int ) -> int: '''simple docstring''' if not isinstance(A_ , A_ ): raise TypeError("Parameter number must be int" ) if number < 0: raise ValueError("Parameter number must be greater than or equal to 0" ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(A_ ) ) def _lowerCamelCase ( A_ : int = 6_0 , A_ : int = 1_0_0_0_0_0_0 ) -> int: '''simple docstring''' if not isinstance(A_ , A_ ) or not isinstance(A_ , A_ ): raise TypeError("Parameters chain_length and number_limit must be int" ) if chain_length <= 0 or number_limit <= 0: raise ValueError( "Parameters chain_length and number_limit must be greater than 0" ) # the counter for the chains with the exact desired length UpperCamelCase__ : str =0 # the cached sizes of the previous chains UpperCamelCase__ : dict[int, int] ={} for start_chain_element in range(1 , A_ ): # The temporary set will contain the elements of the chain UpperCamelCase__ : Any =set() UpperCamelCase__ : Optional[Any] =0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. UpperCamelCase__ : str =start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(A_ ) chain_set_length += 1 UpperCamelCase__ : Tuple =digit_factorial_sum(A_ ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] UpperCamelCase__ : List[str] =chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
582
1
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class A( unittest.TestCase ): """simple docstring""" def _UpperCamelCase( self ) -> Optional[Any]: """simple docstring""" _UpperCamelCase :int = '''laion/clap-htsat-unfused''' _UpperCamelCase :Optional[int] = tempfile.mkdtemp() def _UpperCamelCase( self , **SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: """simple docstring""" return RobertaTokenizer.from_pretrained(self.checkpoint , **SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase( self , **SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" return ClapFeatureExtractor.from_pretrained(self.checkpoint , **SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase( self ) -> int: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _UpperCamelCase( self ) -> Optional[int]: """simple docstring""" _UpperCamelCase :Union[str, Any] = self.get_tokenizer() _UpperCamelCase :Union[str, Any] = self.get_feature_extractor() _UpperCamelCase :Optional[Any] = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) processor.save_pretrained(self.tmpdirname ) _UpperCamelCase :Any = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , SCREAMING_SNAKE_CASE__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase( self ) -> List[str]: """simple docstring""" _UpperCamelCase :List[str] = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _UpperCamelCase :Optional[int] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) _UpperCamelCase :Dict = self.get_feature_extractor(do_normalize=SCREAMING_SNAKE_CASE__ , padding_value=1.0 ) _UpperCamelCase :Optional[int] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=SCREAMING_SNAKE_CASE__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , SCREAMING_SNAKE_CASE__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase :List[Any] = self.get_feature_extractor() _UpperCamelCase :List[str] = self.get_tokenizer() _UpperCamelCase :Union[str, Any] = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) _UpperCamelCase :Dict = floats_list((3, 10_00) ) _UpperCamelCase :Optional[int] = feature_extractor(SCREAMING_SNAKE_CASE__ , return_tensors='''np''' ) _UpperCamelCase :Optional[int] = processor(audios=SCREAMING_SNAKE_CASE__ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _UpperCamelCase( self ) -> Tuple: """simple docstring""" _UpperCamelCase :Union[str, Any] = self.get_feature_extractor() _UpperCamelCase :Optional[int] = self.get_tokenizer() _UpperCamelCase :List[str] = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) _UpperCamelCase :List[Any] = '''This is a test string''' _UpperCamelCase :str = processor(text=SCREAMING_SNAKE_CASE__ ) _UpperCamelCase :Optional[Any] = tokenizer(SCREAMING_SNAKE_CASE__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _UpperCamelCase( self ) -> Optional[Any]: """simple docstring""" _UpperCamelCase :int = self.get_feature_extractor() _UpperCamelCase :Union[str, Any] = self.get_tokenizer() _UpperCamelCase :Tuple = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) _UpperCamelCase :Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _UpperCamelCase :str = processor.batch_decode(SCREAMING_SNAKE_CASE__ ) _UpperCamelCase :Dict = tokenizer.batch_decode(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase( self ) -> Optional[int]: """simple docstring""" _UpperCamelCase :Union[str, Any] = self.get_feature_extractor() _UpperCamelCase :Any = self.get_tokenizer() _UpperCamelCase :Optional[int] = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , )
355
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() UpperCamelCase__ :Optional[int] = logging.get_logger(__name__) UpperCamelCase__ :int = { """post_extract_proj""": """feature_projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.upsample.0""": """encoder.upsample.projection""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """layer_norm""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } def A_ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Tuple: for attribute in key.split('''.''' ): _UpperCamelCase :Any = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _UpperCamelCase :Any = getattr(snake_case__ , snake_case__ ).shape else: _UpperCamelCase :Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( f"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" f" {value.shape} for {full_name}" ) if weight_type == "weight": _UpperCamelCase :str = value elif weight_type == "weight_g": _UpperCamelCase :Dict = value elif weight_type == "weight_v": _UpperCamelCase :Optional[Any] = value elif weight_type == "bias": _UpperCamelCase :str = value else: _UpperCamelCase :int = value logger.info(f"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def A_ ( snake_case__ , snake_case__ , snake_case__ ) -> Tuple: _UpperCamelCase :Optional[int] = [] _UpperCamelCase :List[str] = fairseq_model.state_dict() _UpperCamelCase :str = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _UpperCamelCase :Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == '''group''' , ) _UpperCamelCase :Dict = True else: for key, mapped_key in MAPPING.items(): _UpperCamelCase :Optional[int] = '''sew.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _UpperCamelCase :List[str] = True if "*" in mapped_key: _UpperCamelCase :List[str] = name.split(snake_case__ )[0].split('''.''' )[-2] _UpperCamelCase :Tuple = mapped_key.replace('''*''' , snake_case__ ) if "weight_g" in name: _UpperCamelCase :List[Any] = '''weight_g''' elif "weight_v" in name: _UpperCamelCase :Union[str, Any] = '''weight_v''' elif "weight" in name: _UpperCamelCase :List[Any] = '''weight''' elif "bias" in name: _UpperCamelCase :List[Any] = '''bias''' else: _UpperCamelCase :List[Any] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(f"Unused weights: {unused_weights}" ) def A_ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[Any]: _UpperCamelCase :Optional[int] = full_name.split('''conv_layers.''' )[-1] _UpperCamelCase :Optional[int] = name.split('''.''' ) _UpperCamelCase :Optional[Any] = int(items[0] ) _UpperCamelCase :List[str] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _UpperCamelCase :Optional[int] = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _UpperCamelCase :Optional[Any] = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _UpperCamelCase :int = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _UpperCamelCase :Dict = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def A_ ( snake_case__ , snake_case__ ) -> List[str]: _UpperCamelCase :str = SEWConfig() if is_finetuned: _UpperCamelCase :Optional[int] = model.wav_encoder.wav_model.cfg else: _UpperCamelCase :Dict = model.cfg _UpperCamelCase :Dict = fs_config.conv_bias _UpperCamelCase :int = eval(fs_config.conv_feature_layers ) _UpperCamelCase :List[Any] = [x[0] for x in conv_layers] _UpperCamelCase :Optional[int] = [x[1] for x in conv_layers] _UpperCamelCase :Optional[int] = [x[2] for x in conv_layers] _UpperCamelCase :str = '''gelu''' _UpperCamelCase :Optional[int] = '''layer''' if fs_config.extractor_mode == '''layer_norm''' else '''group''' _UpperCamelCase :List[Any] = 0.0 _UpperCamelCase :Optional[int] = fs_config.activation_fn.name _UpperCamelCase :str = fs_config.encoder_embed_dim _UpperCamelCase :Dict = 0.02 _UpperCamelCase :Optional[int] = fs_config.encoder_ffn_embed_dim _UpperCamelCase :str = 1E-5 _UpperCamelCase :int = fs_config.encoder_layerdrop _UpperCamelCase :Union[str, Any] = fs_config.encoder_attention_heads _UpperCamelCase :List[str] = fs_config.conv_pos_groups _UpperCamelCase :List[Any] = fs_config.conv_pos _UpperCamelCase :List[str] = len(snake_case__ ) _UpperCamelCase :Optional[int] = fs_config.encoder_layers _UpperCamelCase :Optional[int] = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _UpperCamelCase :List[Any] = model.cfg _UpperCamelCase :List[Any] = fs_config.final_dropout _UpperCamelCase :Dict = fs_config.layerdrop _UpperCamelCase :Any = fs_config.activation_dropout _UpperCamelCase :List[str] = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _UpperCamelCase :Optional[Any] = fs_config.attention_dropout _UpperCamelCase :List[Any] = fs_config.dropout_input _UpperCamelCase :Dict = fs_config.dropout _UpperCamelCase :int = fs_config.mask_channel_length _UpperCamelCase :Tuple = fs_config.mask_channel_prob _UpperCamelCase :int = fs_config.mask_length _UpperCamelCase :Dict = fs_config.mask_prob _UpperCamelCase :List[Any] = '''Wav2Vec2FeatureExtractor''' _UpperCamelCase :Optional[Any] = '''Wav2Vec2CTCTokenizer''' return config @torch.no_grad() def A_ ( snake_case__ , snake_case__ , snake_case__=None , snake_case__=None , snake_case__=True ) -> int: if is_finetuned: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase :List[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase :str = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _UpperCamelCase :Any = SEWConfig.from_pretrained(snake_case__ ) else: _UpperCamelCase :List[str] = convert_config(model[0] , snake_case__ ) _UpperCamelCase :List[str] = model[0].eval() _UpperCamelCase :Optional[Any] = True if config.feat_extract_norm == '''layer''' else False _UpperCamelCase :str = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) if is_finetuned: if dict_path: _UpperCamelCase :int = Dictionary.load(snake_case__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _UpperCamelCase :List[Any] = target_dict.pad_index _UpperCamelCase :str = target_dict.bos_index _UpperCamelCase :int = target_dict.pad_index _UpperCamelCase :Dict = target_dict.bos_index _UpperCamelCase :int = target_dict.eos_index _UpperCamelCase :str = len(target_dict.symbols ) _UpperCamelCase :Optional[int] = os.path.join(snake_case__ , '''vocab.json''' ) if not os.path.isdir(snake_case__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(snake_case__ ) ) return os.makedirs(snake_case__ , exist_ok=snake_case__ ) with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , snake_case__ ) _UpperCamelCase :int = WavaVecaCTCTokenizer( snake_case__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=snake_case__ , ) _UpperCamelCase :Dict = WavaVecaProcessor(feature_extractor=snake_case__ , tokenizer=snake_case__ ) processor.save_pretrained(snake_case__ ) _UpperCamelCase :Tuple = SEWForCTC(snake_case__ ) else: _UpperCamelCase :str = SEWModel(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) recursively_load_weights(snake_case__ , snake_case__ , snake_case__ ) hf_model.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase__ :List[str] = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--is_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) UpperCamelCase__ :Tuple = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
355
1
'''simple docstring''' def __a ( lowerCAmelCase__ : list[list[int]] , lowerCAmelCase__ : int , lowerCAmelCase__ : int , lowerCAmelCase__ : set ): a__ , a__ : Dict = len(lowerCAmelCase__ ), len(grid[0] ) if ( min(lowerCAmelCase__ , lowerCAmelCase__ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) a__ : List[str] = 0 count += depth_first_search(lowerCAmelCase__ , row + 1 , lowerCAmelCase__ , lowerCAmelCase__ ) count += depth_first_search(lowerCAmelCase__ , row - 1 , lowerCAmelCase__ , lowerCAmelCase__ ) count += depth_first_search(lowerCAmelCase__ , lowerCAmelCase__ , col + 1 , lowerCAmelCase__ ) count += depth_first_search(lowerCAmelCase__ , lowerCAmelCase__ , col - 1 , lowerCAmelCase__ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
340
'''simple docstring''' class lowerCAmelCase__ : """simple docstring""" def __init__( self : Optional[Any] , A__ : list[int] ) -> None: '''simple docstring''' a__ : Union[str, Any] = len(A__ ) a__ : Tuple = [0] * len_array if len_array > 0: a__ : Dict = array[0] for i in range(1 , A__ ): a__ : Optional[Any] = self.prefix_sum[i - 1] + array[i] def __lowerCAmelCase ( self : int , A__ : int , A__ : int ) -> int: '''simple docstring''' if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def __lowerCAmelCase ( self : Tuple , A__ : int ) -> bool: '''simple docstring''' a__ : Tuple = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(A__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
340
1
'''simple docstring''' import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = OrderedDict( [ ('''align''', '''EfficientNetImageProcessor'''), ('''beit''', '''BeitImageProcessor'''), ('''bit''', '''BitImageProcessor'''), ('''blip''', '''BlipImageProcessor'''), ('''blip-2''', '''BlipImageProcessor'''), ('''bridgetower''', '''BridgeTowerImageProcessor'''), ('''chinese_clip''', '''ChineseCLIPImageProcessor'''), ('''clip''', '''CLIPImageProcessor'''), ('''clipseg''', '''ViTImageProcessor'''), ('''conditional_detr''', '''ConditionalDetrImageProcessor'''), ('''convnext''', '''ConvNextImageProcessor'''), ('''convnextv2''', '''ConvNextImageProcessor'''), ('''cvt''', '''ConvNextImageProcessor'''), ('''data2vec-vision''', '''BeitImageProcessor'''), ('''deformable_detr''', '''DeformableDetrImageProcessor'''), ('''deit''', '''DeiTImageProcessor'''), ('''deta''', '''DetaImageProcessor'''), ('''detr''', '''DetrImageProcessor'''), ('''dinat''', '''ViTImageProcessor'''), ('''donut-swin''', '''DonutImageProcessor'''), ('''dpt''', '''DPTImageProcessor'''), ('''efficientformer''', '''EfficientFormerImageProcessor'''), ('''efficientnet''', '''EfficientNetImageProcessor'''), ('''flava''', '''FlavaImageProcessor'''), ('''focalnet''', '''BitImageProcessor'''), ('''git''', '''CLIPImageProcessor'''), ('''glpn''', '''GLPNImageProcessor'''), ('''groupvit''', '''CLIPImageProcessor'''), ('''imagegpt''', '''ImageGPTImageProcessor'''), ('''instructblip''', '''BlipImageProcessor'''), ('''layoutlmv2''', '''LayoutLMv2ImageProcessor'''), ('''layoutlmv3''', '''LayoutLMv3ImageProcessor'''), ('''levit''', '''LevitImageProcessor'''), ('''mask2former''', '''Mask2FormerImageProcessor'''), ('''maskformer''', '''MaskFormerImageProcessor'''), ('''mgp-str''', '''ViTImageProcessor'''), ('''mobilenet_v1''', '''MobileNetV1ImageProcessor'''), ('''mobilenet_v2''', '''MobileNetV2ImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevitv2''', '''MobileViTImageProcessor'''), ('''nat''', '''ViTImageProcessor'''), ('''oneformer''', '''OneFormerImageProcessor'''), ('''owlvit''', '''OwlViTImageProcessor'''), ('''perceiver''', '''PerceiverImageProcessor'''), ('''pix2struct''', '''Pix2StructImageProcessor'''), ('''poolformer''', '''PoolFormerImageProcessor'''), ('''regnet''', '''ConvNextImageProcessor'''), ('''resnet''', '''ConvNextImageProcessor'''), ('''sam''', '''SamImageProcessor'''), ('''segformer''', '''SegformerImageProcessor'''), ('''swiftformer''', '''ViTImageProcessor'''), ('''swin''', '''ViTImageProcessor'''), ('''swin2sr''', '''Swin2SRImageProcessor'''), ('''swinv2''', '''ViTImageProcessor'''), ('''table-transformer''', '''DetrImageProcessor'''), ('''timesformer''', '''VideoMAEImageProcessor'''), ('''tvlt''', '''TvltImageProcessor'''), ('''upernet''', '''SegformerImageProcessor'''), ('''van''', '''ConvNextImageProcessor'''), ('''videomae''', '''VideoMAEImageProcessor'''), ('''vilt''', '''ViltImageProcessor'''), ('''vit''', '''ViTImageProcessor'''), ('''vit_hybrid''', '''ViTHybridImageProcessor'''), ('''vit_mae''', '''ViTImageProcessor'''), ('''vit_msn''', '''ViTImageProcessor'''), ('''xclip''', '''CLIPImageProcessor'''), ('''yolos''', '''YolosImageProcessor'''), ] ) lowerCAmelCase__ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def _A ( A__ ): """simple docstring""" for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: __lowercase = model_type_to_module_name(A__ ) __lowercase = importlib.import_module(F".{module_name}" , '''transformers.models''' ) try: return getattr(A__ , A__ ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(A__ , '''__name__''' , A__ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __lowercase = importlib.import_module('''transformers''' ) if hasattr(A__ , A__ ): return getattr(A__ , A__ ) return None def _A ( A__ , A__ = None , A__ = False , A__ = False , A__ = None , A__ = None , A__ = None , A__ = False , **A__ , ): """simple docstring""" __lowercase = get_file_from_repo( A__ , A__ , cache_dir=A__ , force_download=A__ , resume_download=A__ , proxies=A__ , use_auth_token=A__ , revision=A__ , local_files_only=A__ , ) if resolved_config_file is None: logger.info( '''Could not locate the image processor configuration file, will try to use the model config instead.''' ) return {} with open(A__ , encoding='''utf-8''' ) as reader: return json.load(A__ ) class lowercase_ : """simple docstring""" def __init__( self : int ): raise EnvironmentError( '''AutoImageProcessor is designed to be instantiated ''' '''using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.''' ) @classmethod @replace_list_option_in_docstrings(lowercase__ ) def SCREAMING_SNAKE_CASE ( cls : str ,lowercase__ : str ,**lowercase__ : int ): __lowercase = kwargs.pop('''config''' ,lowercase__ ) __lowercase = kwargs.pop('''trust_remote_code''' ,lowercase__ ) __lowercase = True __lowercase , __lowercase = ImageProcessingMixin.get_image_processor_dict(lowercase__ ,**lowercase__ ) __lowercase = config_dict.get('''image_processor_type''' ,lowercase__ ) __lowercase = None if "AutoImageProcessor" in config_dict.get('''auto_map''' ,{} ): __lowercase = config_dict['''auto_map''']['''AutoImageProcessor'''] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: __lowercase = config_dict.pop('''feature_extractor_type''' ,lowercase__ ) if feature_extractor_class is not None: logger.warning( '''Could not find image processor class in the image processor config or the model config. Loading''' ''' based on pattern matching with the model\'s feature extractor configuration.''' ) __lowercase = feature_extractor_class.replace('''FeatureExtractor''' ,'''ImageProcessor''' ) if "AutoFeatureExtractor" in config_dict.get('''auto_map''' ,{} ): __lowercase = config_dict['''auto_map''']['''AutoFeatureExtractor'''] __lowercase = feature_extractor_auto_map.replace('''FeatureExtractor''' ,'''ImageProcessor''' ) logger.warning( '''Could not find image processor auto map in the image processor config or the model config.''' ''' Loading based on pattern matching with the model\'s feature extractor configuration.''' ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(lowercase__ ,lowercase__ ): __lowercase = AutoConfig.from_pretrained(lowercase__ ,**lowercase__ ) # It could be in `config.image_processor_type`` __lowercase = getattr(lowercase__ ,'''image_processor_type''' ,lowercase__ ) if hasattr(lowercase__ ,'''auto_map''' ) and "AutoImageProcessor" in config.auto_map: __lowercase = config.auto_map['''AutoImageProcessor'''] if image_processor_class is not None: __lowercase = image_processor_class_from_name(lowercase__ ) __lowercase = image_processor_auto_map is not None __lowercase = image_processor_class is not None or type(lowercase__ ) in IMAGE_PROCESSOR_MAPPING __lowercase = resolve_trust_remote_code( lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ ) if has_remote_code and trust_remote_code: __lowercase = get_class_from_dynamic_module( lowercase__ ,lowercase__ ,**lowercase__ ) __lowercase = kwargs.pop('''code_revision''' ,lowercase__ ) if os.path.isdir(lowercase__ ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(lowercase__ ,**lowercase__ ) elif image_processor_class is not None: return image_processor_class.from_dict(lowercase__ ,**lowercase__ ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(lowercase__ ) in IMAGE_PROCESSOR_MAPPING: __lowercase = IMAGE_PROCESSOR_MAPPING[type(lowercase__ )] return image_processor_class.from_dict(lowercase__ ,**lowercase__ ) raise ValueError( F"Unrecognized image processor in {pretrained_model_name_or_path}. Should have a " F"`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following " F"`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}" ) @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : Tuple ,lowercase__ : Any ): IMAGE_PROCESSOR_MAPPING.register(lowercase__ ,lowercase__ )
41
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def __A(lowerCAmelCase , lowerCAmelCase ) -> np.array: """simple docstring""" _UpperCamelCase = F'{sampling_rate}' _UpperCamelCase = """1""" _UpperCamelCase = """f32le""" _UpperCamelCase = [ """ffmpeg""", """-i""", """pipe:0""", """-ac""", ac, """-ar""", ar, """-f""", format_for_conversion, """-hide_banner""", """-loglevel""", """quiet""", """pipe:1""", ] try: with subprocess.Popen(lowerCAmelCase , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _UpperCamelCase = ffmpeg_process.communicate(lowerCAmelCase ) except FileNotFoundError as error: raise ValueError("""ffmpeg was not found but is required to load audio files from filename""" ) from error _UpperCamelCase = output_stream[0] _UpperCamelCase = np.frombuffer(lowerCAmelCase , np.floataa ) if audio.shape[0] == 0: raise ValueError("""Malformed soundfile""" ) return audio def __A(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = "f32le" , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = F'{sampling_rate}' _UpperCamelCase = """1""" if format_for_conversion == "s16le": _UpperCamelCase = 2 elif format_for_conversion == "f32le": _UpperCamelCase = 4 else: raise ValueError(F'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) _UpperCamelCase = platform.system() if system == "Linux": _UpperCamelCase = """alsa""" _UpperCamelCase = """default""" elif system == "Darwin": _UpperCamelCase = """avfoundation""" _UpperCamelCase = """:0""" elif system == "Windows": _UpperCamelCase = """dshow""" _UpperCamelCase = """default""" _UpperCamelCase = [ """ffmpeg""", """-f""", format_, """-i""", input_, """-ac""", ac, """-ar""", ar, """-f""", format_for_conversion, """-fflags""", """nobuffer""", """-hide_banner""", """-loglevel""", """quiet""", """pipe:1""", ] _UpperCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _UpperCamelCase = _ffmpeg_stream(lowerCAmelCase , lowerCAmelCase ) for item in iterator: yield item def __A(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "f32le" , ) -> Any: """simple docstring""" if stream_chunk_s is not None: _UpperCamelCase = stream_chunk_s else: _UpperCamelCase = chunk_length_s _UpperCamelCase = ffmpeg_microphone(lowerCAmelCase , lowerCAmelCase , format_for_conversion=lowerCAmelCase ) if format_for_conversion == "s16le": _UpperCamelCase = np.intaa _UpperCamelCase = 2 elif format_for_conversion == "f32le": _UpperCamelCase = np.floataa _UpperCamelCase = 4 else: raise ValueError(F'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) if stride_length_s is None: _UpperCamelCase = chunk_length_s / 6 _UpperCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowerCAmelCase , (int, float) ): _UpperCamelCase = [stride_length_s, stride_length_s] _UpperCamelCase = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _UpperCamelCase = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _UpperCamelCase = datetime.datetime.now() _UpperCamelCase = datetime.timedelta(seconds=lowerCAmelCase ) for item in chunk_bytes_iter(lowerCAmelCase , lowerCAmelCase , stride=(stride_left, stride_right) , stream=lowerCAmelCase ): # Put everything back in numpy scale _UpperCamelCase = np.frombuffer(item["""raw"""] , dtype=lowerCAmelCase ) _UpperCamelCase = ( item["""stride"""][0] // size_of_sample, item["""stride"""][1] // size_of_sample, ) _UpperCamelCase = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 1_0 * delta: # We're late !! SKIP continue yield item def __A(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = False ) -> str: """simple docstring""" _UpperCamelCase = b"""""" _UpperCamelCase , _UpperCamelCase = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}' ) _UpperCamelCase = 0 for raw in iterator: acc += raw if stream and len(lowerCAmelCase ) < chunk_len: _UpperCamelCase = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowerCAmelCase ) >= chunk_len: # We are flushing the accumulator _UpperCamelCase = (_stride_left, stride_right) _UpperCamelCase = {"""raw""": acc[:chunk_len], """stride""": stride} if stream: _UpperCamelCase = False yield item _UpperCamelCase = stride_left _UpperCamelCase = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowerCAmelCase ) > stride_left: _UpperCamelCase = {"""raw""": acc, """stride""": (_stride_left, 0)} if stream: _UpperCamelCase = False yield item def __A(lowerCAmelCase , lowerCAmelCase ) -> Tuple: """simple docstring""" _UpperCamelCase = 2**2_4 # 16Mo try: with subprocess.Popen(lowerCAmelCase , stdout=subprocess.PIPE , bufsize=lowerCAmelCase ) as ffmpeg_process: while True: _UpperCamelCase = ffmpeg_process.stdout.read(lowerCAmelCase ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("""ffmpeg was not found but is required to stream audio files from filename""" ) from error
612
0
def snake_case_ ( lowercase__ : int ) -> str: '''simple docstring''' if len(__snake_case ) <= 1: return [tuple(__snake_case )] _lowerCAmelCase =[] def generate(lowercase__ : Optional[Any] , lowercase__ : Tuple ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , __snake_case ) for i in range(k - 1 ): if k % 2 == 0: # k is even _lowerCAmelCase , _lowerCAmelCase =arr[k - 1], arr[i] else: # k is odd _lowerCAmelCase , _lowerCAmelCase =arr[k - 1], arr[0] generate(k - 1 , __snake_case ) generate(len(__snake_case ) , __snake_case ) return res if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = input('''Enter numbers separated by a comma:\n''').strip() __SCREAMING_SNAKE_CASE : Dict = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
705
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel __SCREAMING_SNAKE_CASE : List[str] = { '''text_branch''': '''text_model''', '''audio_branch''': '''audio_model.audio_encoder''', '''attn''': '''attention.self''', '''self.proj''': '''output.dense''', '''attention.self_mask''': '''attn_mask''', '''mlp.fc1''': '''intermediate.dense''', '''mlp.fc2''': '''output.dense''', '''norm1''': '''layernorm_before''', '''norm2''': '''layernorm_after''', '''bn0''': '''batch_norm''', } __SCREAMING_SNAKE_CASE : Dict = AutoFeatureExtractor.from_pretrained('''laion/clap-htsat-unfused''', truncation='''rand_trunc''') def snake_case_ ( lowercase__ : Any , lowercase__ : Union[str, Any]=False ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase =create_model( """HTSAT-tiny""" , """roberta""" , lowercase__ , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=lowercase__ , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def snake_case_ ( lowercase__ : str ): '''simple docstring''' _lowerCAmelCase ={} _lowerCAmelCase =r""".*sequential.(\d+).*""" _lowerCAmelCase =r""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _lowerCAmelCase =key.replace(lowercase__ , lowercase__ ) if re.match(lowercase__ , lowercase__ ): # replace sequential layers with list _lowerCAmelCase =re.match(lowercase__ , lowercase__ ).group(1 ) _lowerCAmelCase =key.replace(f"sequential.{sequential_layer}." , f"layers.{int(lowercase__ )//3}.linear." ) elif re.match(lowercase__ , lowercase__ ): _lowerCAmelCase =int(re.match(lowercase__ , lowercase__ ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... _lowerCAmelCase =1 if projecton_layer == 0 else 2 _lowerCAmelCase =key.replace(f"_projection.{projecton_layer}." , f"_projection.linear{transformers_projection_layer}." ) if "audio" and "qkv" in key: # split qkv into query key and value _lowerCAmelCase =value _lowerCAmelCase =mixed_qkv.size(0 ) // 3 _lowerCAmelCase =mixed_qkv[:qkv_dim] _lowerCAmelCase =mixed_qkv[qkv_dim : qkv_dim * 2] _lowerCAmelCase =mixed_qkv[qkv_dim * 2 :] _lowerCAmelCase =query_layer _lowerCAmelCase =key_layer _lowerCAmelCase =value_layer else: _lowerCAmelCase =value return model_state_dict def snake_case_ ( lowercase__ : int , lowercase__ : List[Any] , lowercase__ : Optional[int] , lowercase__ : Optional[Any]=False ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase =init_clap(lowercase__ , enable_fusion=lowercase__ ) clap_model.eval() _lowerCAmelCase =clap_model.state_dict() _lowerCAmelCase =rename_state_dict(lowercase__ ) _lowerCAmelCase =ClapConfig() _lowerCAmelCase =enable_fusion _lowerCAmelCase =ClapModel(lowercase__ ) # ignore the spectrogram embedding layer model.load_state_dict(lowercase__ , strict=lowercase__ ) model.save_pretrained(lowercase__ ) transformers_config.save_pretrained(lowercase__ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument('''--enable_fusion''', action='''store_true''', help='''Whether to enable fusion or not''') __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
149
0
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase : '''simple docstring''' def __init__( self , UpperCamelCase_ , UpperCamelCase_=13 , UpperCamelCase_=32 , UpperCamelCase_=2 , UpperCamelCase_=3 , UpperCamelCase_=16 , UpperCamelCase_=[1, 2, 1] , UpperCamelCase_=[2, 2, 4] , UpperCamelCase_=2 , UpperCamelCase_=2.0 , UpperCamelCase_=True , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.1 , UpperCamelCase_="gelu" , UpperCamelCase_=False , UpperCamelCase_=True , UpperCamelCase_=0.02 , UpperCamelCase_=1E-5 , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=True , UpperCamelCase_=10 , UpperCamelCase_=8 , ): lowercase_ :List[Any] = parent lowercase_ :Union[str, Any] = batch_size lowercase_ :Union[str, Any] = image_size lowercase_ :Any = patch_size lowercase_ :int = num_channels lowercase_ :Dict = embed_dim lowercase_ :int = depths lowercase_ :List[Any] = num_heads lowercase_ :Any = window_size lowercase_ :int = mlp_ratio lowercase_ :Dict = qkv_bias lowercase_ :Dict = hidden_dropout_prob lowercase_ :Optional[Any] = attention_probs_dropout_prob lowercase_ :Optional[Any] = drop_path_rate lowercase_ :str = hidden_act lowercase_ :int = use_absolute_embeddings lowercase_ :List[str] = patch_norm lowercase_ :Dict = layer_norm_eps lowercase_ :Tuple = initializer_range lowercase_ :Optional[int] = is_training lowercase_ :List[Any] = scope lowercase_ :List[Any] = use_labels lowercase_ :str = type_sequence_label_size lowercase_ :int = encoder_stride def UpperCamelCase ( self ): lowercase_ :str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ :Dict = None if self.use_labels: lowercase_ :Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ :List[Any] = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :List[str] = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ :str = model(__UpperCamelCase ) lowercase_ :str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowercase_ :str = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :List[Any] = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ :List[str] = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowercase_ :Dict = 1 lowercase_ :str = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ :Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase_ :Any = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :List[str] = self.type_sequence_label_size lowercase_ :Any = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowercase_ :Optional[int] = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ): lowercase_ :Union[str, Any] = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ :str = config_and_inputs lowercase_ :str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' lowercase : Union[str, Any] =( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase : str =( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) lowercase : str =False lowercase : List[Any] =False lowercase : int =False lowercase : Optional[int] =False def UpperCamelCase ( self ): lowercase_ :str = SwinvaModelTester(self ) lowercase_ :List[Any] = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def UpperCamelCase ( self ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): lowercase_ :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''' ) def UpperCamelCase ( self ): pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''' ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): lowercase_ , lowercase_ :Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ :str = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase_ :Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def UpperCamelCase ( self ): lowercase_ , lowercase_ :int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ :Any = model_class(__UpperCamelCase ) lowercase_ :Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ :int = [*signature.parameters.keys()] lowercase_ :List[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def UpperCamelCase ( self ): lowercase_ , lowercase_ :List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ :Tuple = True for model_class in self.all_model_classes: lowercase_ :Union[str, Any] = True lowercase_ :Tuple = False lowercase_ :List[Any] = True lowercase_ :Dict = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): lowercase_ :int = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) lowercase_ :int = outputs.attentions lowercase_ :Optional[Any] = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowercase_ :str = True lowercase_ :Optional[Any] = config.window_size**2 lowercase_ :Tuple = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): lowercase_ :Dict = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) lowercase_ :Union[str, Any] = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowercase_ :Union[str, Any] = len(__UpperCamelCase ) # Check attention is always last and order is fine lowercase_ :List[str] = True lowercase_ :Optional[int] = True lowercase_ :Union[str, Any] = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): lowercase_ :Optional[int] = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , '''num_hidden_states_types''' ): lowercase_ :Any = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowercase_ :Dict = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) lowercase_ :Dict = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def UpperCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ :Optional[Any] = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): lowercase_ :List[str] = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) lowercase_ :Dict = outputs.hidden_states lowercase_ :Union[str, Any] = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length lowercase_ :Dict = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowercase_ :Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowercase_ :Any = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) lowercase_ , lowercase_ , lowercase_ , lowercase_ :List[str] = reshaped_hidden_states[0].shape lowercase_ :int = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): lowercase_ , lowercase_ :List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ :Optional[int] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowercase_ :List[str] = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ :int = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def UpperCamelCase ( self ): lowercase_ , lowercase_ :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ :Dict = 3 lowercase_ :List[str] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowercase_ :Any = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowercase_ :Dict = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowercase_ :Tuple = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowercase_ :Optional[int] = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ :Tuple = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def UpperCamelCase ( self ): lowercase_ :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def UpperCamelCase ( self ): lowercase_ :List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def UpperCamelCase ( self ): for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ :Optional[Any] = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def UpperCamelCase ( self ): lowercase_ , lowercase_ :int = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ :List[Any] = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: lowercase_ :int = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"Parameter {name} of model {model_class} seems not properly initialized" , ) @require_vision @require_torch class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase ( self ): return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): lowercase_ :List[Any] = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ).to( __UpperCamelCase ) lowercase_ :Any = self.default_image_processor lowercase_ :Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowercase_ :Any = image_processor(images=__UpperCamelCase , return_tensors='''pt''' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): lowercase_ :int = model(**__UpperCamelCase ) # verify the logits lowercase_ :Any = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) lowercase_ :List[str] = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
257
from __future__ import annotations from collections import namedtuple def __lowerCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> tuple: _UpperCAmelCase = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
684
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[int] , __a : Optional[int] , __a : List[str]=13 , __a : Optional[Any]=3 , __a : Optional[Any]=224 , __a : str=30 , __a : Optional[Any]=400 , __a : List[Any]=True , __a : Any=None , __a : Union[str, Any]=True , __a : Dict=[0.5, 0.5, 0.5] , __a : str=[0.5, 0.5, 0.5] , ) -> Tuple: _UpperCamelCase : Any = size if size is not None else {"height": 18, "width": 18} _UpperCamelCase : Dict = parent _UpperCamelCase : Optional[Any] = batch_size _UpperCamelCase : List[Any] = num_channels _UpperCamelCase : Union[str, Any] = image_size _UpperCamelCase : Optional[int] = min_resolution _UpperCamelCase : int = max_resolution _UpperCamelCase : str = do_resize _UpperCamelCase : Tuple = size _UpperCamelCase : int = do_normalize _UpperCamelCase : int = image_mean _UpperCamelCase : Tuple = image_std def __SCREAMING_SNAKE_CASE ( self : str ) -> str: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :str = ViTImageProcessor if is_vision_available() else None def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Dict: _UpperCamelCase : str = EfficientFormerImageProcessorTester(self ) @property def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[str]: return self.image_proc_tester.prepare_image_processor_dict() def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> str: _UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__a , "image_mean" ) ) self.assertTrue(hasattr(__a , "image_std" ) ) self.assertTrue(hasattr(__a , "do_normalize" ) ) self.assertTrue(hasattr(__a , "do_resize" ) ) self.assertTrue(hasattr(__a , "size" ) ) def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: pass def __SCREAMING_SNAKE_CASE ( self : str ) -> Dict: # Initialize image_processor _UpperCamelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCamelCase : List[str] = prepare_image_inputs(self.image_proc_tester , equal_resolution=__a ) for image in image_inputs: self.assertIsInstance(__a , Image.Image ) # Test not batched input _UpperCamelCase : Union[str, Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _UpperCamelCase : Any = image_processor(__a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Dict: # Initialize image_processor _UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCamelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=__a , numpify=__a ) for image in image_inputs: self.assertIsInstance(__a , np.ndarray ) # Test not batched input _UpperCamelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _UpperCamelCase : Optional[Any] = image_processor(__a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Any: # Initialize image_processor _UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCamelCase : Optional[int] = prepare_image_inputs(self.image_proc_tester , equal_resolution=__a , torchify=__a ) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor ) # Test not batched input _UpperCamelCase : List[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _UpperCamelCase : Tuple = image_processor(__a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
708
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCamelCase__ = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCamelCase__ = [ord(letter) for letter in string.ascii_lowercase] lowerCamelCase__ = {ord(char) for char in VALID_CHARS} lowerCamelCase__ = ["the", "be", "to", "of", "and", "in", "that", "have"] def lowercase__ ( lowercase_ ,lowercase_ ) -> str | None: """simple docstring""" _UpperCamelCase : str = "" _UpperCamelCase : int _UpperCamelCase : int _UpperCamelCase : int for keychar, cipherchar in zip(cycle(lowercase_ ) ,lowercase_ ): _UpperCamelCase : Dict = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowercase_ ) return decoded def lowercase__ ( lowercase_ ) -> list[str]: """simple docstring""" _UpperCamelCase : list[str] = [] for key in product(lowercase_ ,repeat=3 ): _UpperCamelCase : int = try_key(lowercase_ ,lowercase_ ) if encoded is not None: possibles.append(lowercase_ ) return possibles def lowercase__ ( lowercase_ ,lowercase_ ) -> list[str]: """simple docstring""" return [possible for possible in possibles if common_word in possible.lower()] def lowercase__ ( lowercase_ = "p059_cipher.txt" ) -> int: """simple docstring""" _UpperCamelCase : list[int] _UpperCamelCase : list[str] _UpperCamelCase : str _UpperCamelCase : str _UpperCamelCase : str = Path(lowercase_ ).parent.joinpath(lowercase_ ).read_text(encoding="utf-8" ) _UpperCamelCase : Optional[Any] = [int(lowercase_ ) for number in data.strip().split("," )] _UpperCamelCase : List[str] = filter_valid_chars(lowercase_ ) for common_word in COMMON_WORDS: _UpperCamelCase : Union[str, Any] = filter_common_word(lowercase_ ,lowercase_ ) if len(lowercase_ ) == 1: break _UpperCamelCase : Union[str, Any] = possibles[0] return sum(ord(lowercase_ ) for char in decoded_text ) if __name__ == "__main__": print(f"""{solution() = }""")
51
0
import random def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase = False ): '''simple docstring''' UpperCAmelCase_ : dict = {i: [] for i in range(_lowercase )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(_lowercase ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(_lowercase ): for j in range(i + 1 , _lowercase ): if random.random() < probability: graph[i].append(_lowercase ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(_lowercase ) return graph def lowerCamelCase__ ( _lowercase ): '''simple docstring''' return { i: [j for j in range(_lowercase ) if i != j] for i in range(_lowercase ) } if __name__ == "__main__": import doctest doctest.testmod()
30
def _lowercase ( __lowerCamelCase : Optional[int] ,__lowerCamelCase : Dict ,__lowerCamelCase : str ,__lowerCamelCase : Optional[int] ) -> List[str]: '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: UpperCamelCase__ : Optional[Any] = mf_knapsack(i - 1 ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) else: UpperCamelCase__ : Optional[int] = max( mf_knapsack(i - 1 ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) ,mf_knapsack(i - 1 ,__lowerCamelCase ,__lowerCamelCase ,j - wt[i - 1] ) + val[i - 1] ,) UpperCamelCase__ : Optional[int] = val return f[i][j] def _lowercase ( __lowerCamelCase : Optional[int] ,__lowerCamelCase : Tuple ,__lowerCamelCase : List[Any] ,__lowerCamelCase : str ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Tuple = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 ,n + 1 ): for w_ in range(1 ,w + 1 ): if wt[i - 1] <= w_: UpperCamelCase__ : Tuple = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] ,dp[i - 1][w_] ) else: UpperCamelCase__ : Optional[Any] = dp[i - 1][w_] return dp[n][w_], dp def _lowercase ( __lowerCamelCase : int ,__lowerCamelCase : list ,__lowerCamelCase : list ) -> List[Any]: '''simple docstring''' if not (isinstance(__lowerCamelCase ,(list, tuple) ) and isinstance(__lowerCamelCase ,(list, tuple) )): raise ValueError( '''Both the weights and values vectors must be either lists or tuples''' ) UpperCamelCase__ : int = len(__lowerCamelCase ) if num_items != len(__lowerCamelCase ): UpperCamelCase__ : int = ( '''The number of weights must be the same as the number of values.\n''' F'But got {num_items} weights and {len(__lowerCamelCase )} values' ) raise ValueError(__lowerCamelCase ) for i in range(__lowerCamelCase ): if not isinstance(wt[i] ,__lowerCamelCase ): UpperCamelCase__ : Tuple = ( '''All weights must be integers but got weight of ''' F'type {type(wt[i] )} at index {i}' ) raise TypeError(__lowerCamelCase ) UpperCamelCase__ ,UpperCamelCase__ : Tuple = knapsack(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) UpperCamelCase__ : set = set() _construct_solution(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) return optimal_val, example_optional_set def _lowercase ( __lowerCamelCase : list ,__lowerCamelCase : list ,__lowerCamelCase : int ,__lowerCamelCase : int ,__lowerCamelCase : set ) -> Dict: '''simple docstring''' if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(__lowerCamelCase ,__lowerCamelCase ,i - 1 ,__lowerCamelCase ,__lowerCamelCase ) else: optimal_set.add(__lowerCamelCase ) _construct_solution(__lowerCamelCase ,__lowerCamelCase ,i - 1 ,j - wt[i - 1] ,__lowerCamelCase ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Tuple = [3, 2, 4, 4] _SCREAMING_SNAKE_CASE : Any = [4, 3, 2, 3] _SCREAMING_SNAKE_CASE : int = 4 _SCREAMING_SNAKE_CASE : int = 6 _SCREAMING_SNAKE_CASE : Optional[int] = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE : Dict = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE : Any = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("""optimal_value = """, optimal_solution) print("""An optimal subset corresponding to the optimal value""", optimal_subset)
344
0
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline UpperCAmelCase__ = 'path-to-your-trained-model' UpperCAmelCase__ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('cuda') UpperCAmelCase__ = 'A photo of sks dog in a bucket' UpperCAmelCase__ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('dog-bucket.png')
430
"""simple docstring""" from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def _UpperCAmelCase ( __lowerCamelCase : int ) -> Tuple: # A local function to see if a dot lands in the circle. def is_in_circle(__lowerCamelCase : float , __lowerCamelCase : float ) -> bool: _snake_case = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle _snake_case = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(__lowerCamelCase ) ) # The ratio of the area for circle to square is pi/4. _snake_case = proportion * 4 print(f'''The estimated value of pi is {pi_estimate}''' ) print(f'''The numpy value of pi is {pi}''' ) print(f'''The total error is {abs(pi - pi_estimate )}''' ) def _UpperCAmelCase ( __lowerCamelCase : int , __lowerCamelCase : Callable[[float], float] , __lowerCamelCase : float = 0.0 , __lowerCamelCase : float = 1.0 , ) -> float: return mean( function_to_integrate(uniform(__lowerCamelCase , __lowerCamelCase ) ) for _ in range(__lowerCamelCase ) ) * (max_value - min_value) def _UpperCAmelCase ( __lowerCamelCase : int , __lowerCamelCase : float = 0.0 , __lowerCamelCase : float = 1.0 ) -> None: def identity_function(__lowerCamelCase : float ) -> float: return x _snake_case = area_under_curve_estimator( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) _snake_case = (max_value * max_value - min_value * min_value) / 2 print('''******************''' ) print(f'''Estimating area under y=x where x varies from {min_value} to {max_value}''' ) print(f'''Estimated value is {estimated_value}''' ) print(f'''Expected value is {expected_value}''' ) print(f'''Total error is {abs(estimated_value - expected_value )}''' ) print('''******************''' ) def _UpperCAmelCase ( __lowerCamelCase : int ) -> None: def function_to_integrate(__lowerCamelCase : float ) -> float: return sqrt(4.0 - x * x ) _snake_case = area_under_curve_estimator( __lowerCamelCase , __lowerCamelCase , 0.0 , 2.0 ) print('''******************''' ) print('''Estimating pi using area_under_curve_estimator''' ) print(f'''Estimated value is {estimated_value}''' ) print(f'''Expected value is {pi}''' ) print(f'''Total error is {abs(estimated_value - pi )}''' ) print('''******************''' ) if __name__ == "__main__": import doctest doctest.testmod()
430
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __snake_case ( unittest.TestCase ): '''simple docstring''' def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=7 , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=1_0 , __SCREAMING_SNAKE_CASE=1_8 , __SCREAMING_SNAKE_CASE=3_0 , __SCREAMING_SNAKE_CASE=4_0_0 , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=[0.5, 0.5, 0.5] , __SCREAMING_SNAKE_CASE=[0.5, 0.5, 0.5] , __SCREAMING_SNAKE_CASE=None , ): snake_case__ : Union[str, Any] = size if size is not None else {"""shortest_edge""": 1_8} snake_case__ : Tuple = crop_size if crop_size is not None else {"""height""": 1_8, """width""": 1_8} snake_case__ : Dict = parent snake_case__ : str = batch_size snake_case__ : str = num_channels snake_case__ : Optional[int] = num_frames snake_case__ : Tuple = image_size snake_case__ : List[str] = min_resolution snake_case__ : int = max_resolution snake_case__ : str = do_resize snake_case__ : Union[str, Any] = size snake_case__ : str = do_normalize snake_case__ : int = image_mean snake_case__ : List[str] = image_std snake_case__ : Any = crop_size def __UpperCamelCase ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __snake_case ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase ( self ): snake_case__ : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase ( self ): snake_case__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """image_mean""" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """image_std""" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """do_normalize""" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """do_resize""" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """do_center_crop""" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , """size""" ) ) def __UpperCamelCase ( self ): snake_case__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 1_8} ) self.assertEqual(image_processor.crop_size , {"""height""": 1_8, """width""": 1_8} ) snake_case__ : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 4_2} ) self.assertEqual(image_processor.crop_size , {"""height""": 8_4, """width""": 8_4} ) def __UpperCamelCase ( self ): # Initialize image_processing snake_case__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos snake_case__ : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE ) for video in video_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input snake_case__ : Any = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched snake_case__ : List[str] = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __UpperCamelCase ( self ): # Initialize image_processing snake_case__ : int = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case__ : int = prepare_video_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE , numpify=__SCREAMING_SNAKE_CASE ) for video in video_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input snake_case__ : Optional[Any] = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched snake_case__ : Any = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __UpperCamelCase ( self ): # Initialize image_processing snake_case__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case__ : int = prepare_video_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE , torchify=__SCREAMING_SNAKE_CASE ) for video in video_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input snake_case__ : Any = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched snake_case__ : Tuple = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
38
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A_ : int = logging.get_logger(__name__) A_ : Dict = { "google/bit-50": "https://huggingface.co/google/bit-50/resolve/main/config.json", } class __snake_case ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase__ = '''bit''' lowerCamelCase__ = ['''preactivation''', '''bottleneck'''] lowerCamelCase__ = ['''SAME''', '''VALID'''] def __init__( self , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=6_4 , __SCREAMING_SNAKE_CASE=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , __SCREAMING_SNAKE_CASE=[3, 4, 6, 3] , __SCREAMING_SNAKE_CASE="preactivation" , __SCREAMING_SNAKE_CASE="relu" , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=3_2 , __SCREAMING_SNAKE_CASE=0.0 , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=3_2 , __SCREAMING_SNAKE_CASE=1 , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , **__SCREAMING_SNAKE_CASE , ): super().__init__(**__SCREAMING_SNAKE_CASE ) if layer_type not in self.layer_types: raise ValueError(f"layer_type={layer_type} is not one of {','.join(self.layer_types )}" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case__ : Tuple = global_padding.upper() else: raise ValueError(f"Padding strategy {global_padding} not supported" ) snake_case__ : List[str] = num_channels snake_case__ : Tuple = embedding_size snake_case__ : str = hidden_sizes snake_case__ : Optional[Any] = depths snake_case__ : List[Any] = layer_type snake_case__ : Dict = hidden_act snake_case__ : Union[str, Any] = global_padding snake_case__ : List[str] = num_groups snake_case__ : str = drop_path_rate snake_case__ : List[Any] = embedding_dynamic_padding snake_case__ : List[str] = output_stride snake_case__ : Dict = width_factor snake_case__ : List[str] = ["""stem"""] + [f"stage{idx}" for idx in range(1 , len(__SCREAMING_SNAKE_CASE ) + 1 )] snake_case__ , snake_case__ : Dict = get_aligned_output_features_output_indices( out_features=__SCREAMING_SNAKE_CASE , out_indices=__SCREAMING_SNAKE_CASE , stage_names=self.stage_names )
38
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowercase = { '''configuration_pix2struct''': [ '''PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Pix2StructConfig''', '''Pix2StructTextConfig''', '''Pix2StructVisionConfig''', ], '''processing_pix2struct''': ['''Pix2StructProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''Pix2StructImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Pix2StructPreTrainedModel''', '''Pix2StructForConditionalGeneration''', '''Pix2StructVisionModel''', '''Pix2StructTextModel''', ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
709
'''simple docstring''' from typing import Dict, Optional import numpy as np import datasets __lowercase = ''' IoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union between the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation, the mean IoU of the image is calculated by taking the IoU of each class and averaging them. ''' __lowercase = ''' Args: predictions (`List[ndarray]`): List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. references (`List[ndarray]`): List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. num_labels (`int`): Number of classes (categories). ignore_index (`int`): Index that will be ignored during evaluation. nan_to_num (`int`, *optional*): If specified, NaN values will be replaced by the number defined by the user. label_map (`dict`, *optional*): If specified, dictionary mapping old label indices to new label indices. reduce_labels (`bool`, *optional*, defaults to `False`): Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background, and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255. Returns: `Dict[str, float | ndarray]` comprising various elements: - *mean_iou* (`float`): Mean Intersection-over-Union (IoU averaged over all categories). - *mean_accuracy* (`float`): Mean accuracy (averaged over all categories). - *overall_accuracy* (`float`): Overall accuracy on all images. - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`): Per category accuracy. - *per_category_iou* (`ndarray` of shape `(num_labels,)`): Per category IoU. Examples: >>> import numpy as np >>> mean_iou = datasets.load_metric("mean_iou") >>> # suppose one has 3 different segmentation maps predicted >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]]) >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]]) >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]]) >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]]) >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]]) >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]]) >>> predicted = [predicted_1, predicted_2, predicted_3] >>> ground_truth = [actual_1, actual_2, actual_3] >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False) >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mean_iou\': 0.47750000000000004, \'mean_accuracy\': 0.5916666666666666, \'overall_accuracy\': 0.5263157894736842, \'per_category_iou\': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), \'per_category_accuracy\': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])} ''' __lowercase = '''\ @software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020, author = {{MMSegmentation Contributors}}, license = {Apache-2.0}, month = {7}, title = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}}, url = {https://github.com/open-mmlab/mmsegmentation}, year = {2020} }''' def snake_case__ ( _A: Optional[Any] , _A: int , _A: int , _A: bool , _A: Optional[Dict[int, int]] = None , _A: bool = False , ) -> List[Any]: '''simple docstring''' if label_map is not None: for old_id, new_id in label_map.items(): lowerCAmelCase = new_id # turn into Numpy arrays lowerCAmelCase = np.array(_A ) lowerCAmelCase = np.array(_A ) if reduce_labels: lowerCAmelCase = 255 lowerCAmelCase = label - 1 lowerCAmelCase = 255 lowerCAmelCase = label != ignore_index lowerCAmelCase = np.not_equal(_A , _A ) lowerCAmelCase = pred_label[mask] lowerCAmelCase = np.array(_A )[mask] lowerCAmelCase = pred_label[pred_label == label] lowerCAmelCase = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] lowerCAmelCase = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] lowerCAmelCase = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] lowerCAmelCase = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def snake_case__ ( _A: Union[str, Any] , _A: Any , _A: Union[str, Any] , _A: bool , _A: Optional[Dict[int, int]] = None , _A: bool = False , ) -> int: '''simple docstring''' lowerCAmelCase = np.zeros((num_labels,) , dtype=np.floataa ) lowerCAmelCase = np.zeros((num_labels,) , dtype=np.floataa ) lowerCAmelCase = np.zeros((num_labels,) , dtype=np.floataa ) lowerCAmelCase = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(_A , _A ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = intersect_and_union( _A , _A , _A , _A , _A , _A ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def snake_case__ ( _A: Union[str, Any] , _A: int , _A: Dict , _A: bool , _A: Optional[int] = None , _A: Optional[Dict[int, int]] = None , _A: bool = False , ) -> Optional[int]: '''simple docstring''' lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = total_intersect_and_union( _A , _A , _A , _A , _A , _A ) # compute metrics lowerCAmelCase = {} lowerCAmelCase = total_area_intersect.sum() / total_area_label.sum() lowerCAmelCase = total_area_intersect / total_area_union lowerCAmelCase = total_area_intersect / total_area_label lowerCAmelCase = np.nanmean(_A ) lowerCAmelCase = np.nanmean(_A ) lowerCAmelCase = all_acc lowerCAmelCase = iou lowerCAmelCase = acc if nan_to_num is not None: lowerCAmelCase = {metric: np.nan_to_num(_A , nan=_A ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__( datasets.Metric ): '''simple docstring''' def a_ ( self): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { """predictions""": datasets.Sequence(datasets.Sequence(datasets.Value("""uint16"""))), """references""": datasets.Sequence(datasets.Sequence(datasets.Value("""uint16"""))), }) , reference_urls=[ """https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py""" ] , ) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = False , ): """simple docstring""" lowerCAmelCase = mean_iou( results=__lowerCAmelCase , gt_seg_maps=__lowerCAmelCase , num_labels=__lowerCAmelCase , ignore_index=__lowerCAmelCase , nan_to_num=__lowerCAmelCase , label_map=__lowerCAmelCase , reduce_labels=__lowerCAmelCase , ) return iou_result
605
0
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCamelCase__ ( __lowerCamelCase , unittest.TestCase ): a__ : Dict = DanceDiffusionPipeline a__ : int = UNCONDITIONAL_AUDIO_GENERATION_PARAMS a__ : Any = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } a__ : List[str] = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS a__ : Tuple = False a__ : Optional[int] = False def __lowercase( self : Optional[Any] ) -> List[str]: torch.manual_seed(0 ) UpperCamelCase__ : List[str] = UNetaDModel( block_out_channels=(32, 32, 64), extra_in_channels=16, sample_size=5_12, sample_rate=1_60_00, in_channels=2, out_channels=2, flip_sin_to_cos=__lowerCamelCase, use_timestep_embedding=__lowerCamelCase, time_embedding_type='''fourier''', mid_block_type='''UNetMidBlock1D''', down_block_types=('''DownBlock1DNoSkip''', '''DownBlock1D''', '''AttnDownBlock1D'''), up_block_types=('''AttnUpBlock1D''', '''UpBlock1D''', '''UpBlock1DNoSkip'''), ) UpperCamelCase__ : List[str] = IPNDMScheduler() UpperCamelCase__ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, } return components def __lowercase( self : Tuple, __lowerCamelCase : Optional[Any], __lowerCamelCase : str=0 ) -> str: if str(__lowerCamelCase ).startswith('''mps''' ): UpperCamelCase__ : Any = torch.manual_seed(__lowerCamelCase ) else: UpperCamelCase__ : Optional[int] = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) UpperCamelCase__ : List[Any] = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 4, } return inputs def __lowercase( self : Any ) -> Optional[Any]: UpperCamelCase__ : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase__ : Optional[Any] = self.get_dummy_components() UpperCamelCase__ : int = DanceDiffusionPipeline(**__lowerCamelCase ) UpperCamelCase__ : List[str] = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCamelCase__ : Tuple = self.get_dummy_inputs(__lowerCamelCase ) UpperCamelCase__ : Any = pipe(**__lowerCamelCase ) UpperCamelCase__ : Optional[Any] = output.audios UpperCamelCase__ : str = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) UpperCamelCase__ : Optional[int] = np.array([-0.7265, 1.0000, -0.8388, 0.1175, 0.9498, -1.0000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def __lowercase( self : Dict ) -> List[str]: return super().test_save_load_local() @skip_mps def __lowercase( self : Optional[int] ) -> int: return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def __lowercase( self : List[Any] ) -> Any: return super().test_save_load_optional_components() @skip_mps def __lowercase( self : Dict ) -> Tuple: return super().test_attention_slicing_forward_pass() def __lowercase( self : Optional[Any] ) -> Tuple: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): def __lowercase( self : Optional[int] ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase( self : Union[str, Any] ) -> List[str]: UpperCamelCase__ : Optional[Any] = torch_device UpperCamelCase__ : List[str] = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' ) UpperCamelCase__ : int = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCamelCase__ : Optional[int] = torch.manual_seed(0 ) UpperCamelCase__ : Optional[Any] = pipe(generator=__lowerCamelCase, num_inference_steps=1_00, audio_length_in_s=4.096 ) UpperCamelCase__ : Any = output.audios UpperCamelCase__ : Union[str, Any] = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCamelCase__ : int = np.array([-0.0192, -0.0231, -0.0318, -0.0059, 0.0002, -0.0020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def __lowercase( self : Dict ) -> Any: UpperCamelCase__ : Dict = torch_device UpperCamelCase__ : int = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''', torch_dtype=torch.floataa ) UpperCamelCase__ : Dict = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCamelCase__ : Optional[Any] = torch.manual_seed(0 ) UpperCamelCase__ : Optional[Any] = pipe(generator=__lowerCamelCase, num_inference_steps=1_00, audio_length_in_s=4.096 ) UpperCamelCase__ : List[str] = output.audios UpperCamelCase__ : Tuple = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCamelCase__ : Union[str, Any] = np.array([-0.0367, -0.0488, -0.0771, -0.0525, -0.0444, -0.0341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
344
_SCREAMING_SNAKE_CASE : str = 8.3_144_598 def _lowercase ( __lowerCamelCase : float ,__lowerCamelCase : float ) -> float: '''simple docstring''' if temperature < 0: raise Exception('''Temperature cannot be less than 0 K''' ) if molar_mass <= 0: raise Exception('''Molar mass cannot be less than or equal to 0 kg/mol''' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _SCREAMING_SNAKE_CASE : List[str] = 300 _SCREAMING_SNAKE_CASE : Any = 28 _SCREAMING_SNAKE_CASE : Any = rms_speed_of_molecule(temperature, molar_mass) print(F'Vrms of Nitrogen gas at 300 K is {vrms} m/s')
344
1
"""simple docstring""" def a__ ( snake_case__ ) -> int: assert ( isinstance(snake_case__ , snake_case__ ) and number_of_steps > 0 ), F'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 lowerCamelCase , lowerCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): lowerCamelCase , lowerCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
533
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : Optional[Any] = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } lowerCAmelCase : Optional[Any] = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def a__ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> List[Any]: for attribute in key.split(""".""" ): lowerCamelCase = getattr(snake_case__ , snake_case__ ) if weight_type is not None: lowerCamelCase = getattr(snake_case__ , snake_case__ ).shape else: lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": lowerCamelCase = value elif weight_type == "weight_g": lowerCamelCase = value elif weight_type == "weight_v": lowerCamelCase = value elif weight_type == "bias": lowerCamelCase = value else: lowerCamelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def a__ ( snake_case__ , snake_case__ ) -> Tuple: lowerCamelCase = [] lowerCamelCase = fairseq_model.state_dict() lowerCamelCase = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight lowerCamelCase = None for name, value in fairseq_dict.items(): lowerCamelCase = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) lowerCamelCase = True elif name.split(""".""" )[0] == "proj": lowerCamelCase = fairseq_model.proj lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: lowerCamelCase = True if "*" in mapped_key: lowerCamelCase = name.split(snake_case__ )[0].split(""".""" )[-2] lowerCamelCase = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: lowerCamelCase = """weight_g""" elif "weight_v" in name: lowerCamelCase = """weight_v""" elif "bias" in name: lowerCamelCase = """bias""" elif "weight" in name: lowerCamelCase = """weight""" else: lowerCamelCase = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F'Unused weights: {unused_weights}' ) return proj_weight def a__ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[int]: lowerCamelCase = full_name.split("""conv_layers.""" )[-1] lowerCamelCase = name.split(""".""" ) lowerCamelCase = int(items[0] ) lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) lowerCamelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) lowerCamelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(snake_case__ ) def a__ ( snake_case__ ) -> Tuple: lowerCamelCase , lowerCamelCase = emb.weight.shape lowerCamelCase = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) lowerCamelCase = emb.weight.data return lin_layer def a__ ( snake_case__ ) -> Optional[int]: with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: lowerCamelCase = f.readlines() lowerCamelCase = [line.split(""" """ )[0] for line in lines] lowerCamelCase = len(snake_case__ ) lowerCamelCase = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def a__ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) -> Optional[int]: lowerCamelCase = WavaVecaConfig.from_pretrained(snake_case__ ) lowerCamelCase = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) lowerCamelCase , lowerCamelCase , lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) lowerCamelCase = model[0].eval() # set weights for wav2vec2 encoder lowerCamelCase = WavaVecaModel(snake_case__ ) lowerCamelCase = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) lowerCamelCase = SpeechaTextaForCausalLM(snake_case__ ) lowerCamelCase , lowerCamelCase = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) lowerCamelCase = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F'The following keys are missing when loading the decoder weights: {missing_keys}' ) logger.warning(F'The following keys are unexpected when loading the decoder weights: {unexpected_keys}' ) lowerCamelCase = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) lowerCamelCase = False # add projection layer lowerCamelCase = nn.Parameter(projection_layer.weight ) lowerCamelCase = nn.Parameter(projection_layer.bias ) lowerCamelCase = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) lowerCamelCase = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) lowerCamelCase = hf_wavavec.config.to_dict() lowerCamelCase = tokenizer.pad_token_id lowerCamelCase = tokenizer.bos_token_id lowerCamelCase = tokenizer.eos_token_id lowerCamelCase = """speech_to_text_2""" lowerCamelCase = """wav2vec2""" lowerCamelCase = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-large-lv60""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/s2t-small-mustc-en-fr-st""", type=str, help="""Path to hf decoder s2t checkpoint config""", ) parser.add_argument("""--vocab_size""", default=1_0224, type=int, help="""Vocab size of decoder""") parser.add_argument("""--num_decoder_layers""", default=7, type=int, help="""Number of decoder layers""") lowerCAmelCase : str = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
533
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def __a ( a, a=False ): """simple docstring""" _a = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _a = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def __a ( a, a, a=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _a = "" else: _a = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _a = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _a = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _a = in_proj_weight[ : config.hidden_size, : ] _a = in_proj_bias[: config.hidden_size] _a = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _a = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _a = in_proj_weight[ -config.hidden_size :, : ] _a = in_proj_bias[-config.hidden_size :] def __a ( a ): """simple docstring""" _a = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(__a, __a ) def __a ( a, a, a ): """simple docstring""" _a = dct.pop(__a ) _a = val def __a ( ): """simple docstring""" _a = "http://images.cocodataset.org/val2017/000000039769.jpg" _a = Image.open(requests.get(__a, stream=__a ).raw ) return im @torch.no_grad() def __a ( a, a, a=False ): """simple docstring""" _a = BitConfig( global_padding="same", layer_type="bottleneck", depths=(3, 4, 9), out_features=["stage3"], embedding_dynamic_padding=__a, ) _a = ViTHybridConfig(backbone_config=__a, image_size=3_8_4, num_labels=1_0_0_0 ) _a = False # load original model from timm _a = timm.create_model(__a, pretrained=__a ) timm_model.eval() # load state_dict of original model, remove and rename some keys _a = timm_model.state_dict() if base_model: remove_classification_head_(__a ) _a = create_rename_keys(__a, __a ) for src, dest in rename_keys: rename_key(__a, __a, __a ) read_in_q_k_v(__a, __a, __a ) _a = "huggingface/label-files" _a = "imagenet-1k-id2label.json" _a = json.load(open(hf_hub_download(__a, __a, repo_type="dataset" ), "r" ) ) _a = {int(__a ): v for k, v in idalabel.items()} _a = idalabel _a = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": _a = ViTHybridModel(__a ).eval() else: _a = ViTHybridForImageClassification(__a ).eval() model.load_state_dict(__a ) # create image processor _a = create_transform(**resolve_data_config({}, model=__a ) ) _a = transform.transforms _a = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } _a = ViTHybridImageProcessor( do_resize=__a, size={"shortest_edge": timm_transforms[0].size}, resample=pillow_resamplings[timm_transforms[0].interpolation.value], do_center_crop=__a, crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]}, do_normalize=__a, image_mean=timm_transforms[-1].mean.tolist(), image_std=timm_transforms[-1].std.tolist(), ) _a = prepare_img() _a = transform(__a ).unsqueeze(0 ) _a = processor(__a, return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(__a, __a ) # verify logits with torch.no_grad(): _a = model(__a ) _a = outputs.logits print("Predicted class:", logits.argmax(-1 ).item() ) if base_model: _a = timm_model.forward_features(__a ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__a, outputs.pooler_output, atol=1e-3 ) else: _a = timm_model(__a ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__a, outputs.logits, atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(__a ).mkdir(exist_ok=__a ) print(F'Saving model {vit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__a ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(__a ) if push_to_hub: print(F'Pushing model and processor to the hub {vit_name}' ) model.push_to_hub(F'ybelkada/{vit_name}' ) processor.push_to_hub(F'ybelkada/{vit_name}' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_r50_s16_384""", type=str, help="""Name of the hybrid ViT timm model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) __SCREAMING_SNAKE_CASE = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
388
"""simple docstring""" from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract lowerCAmelCase__ = logging.get_logger(__name__) def _lowerCamelCase ( __a, __a, __a ): return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def _lowerCamelCase ( __a, __a, __a ): SCREAMING_SNAKE_CASE_ = to_pil_image(__a ) SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = pil_image.size SCREAMING_SNAKE_CASE_ = pytesseract.image_to_data(__a, lang=__a, output_type='''dict''', config=__a ) SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates SCREAMING_SNAKE_CASE_ = [idx for idx, word in enumerate(__a ) if not word.strip()] SCREAMING_SNAKE_CASE_ = [word for idx, word in enumerate(__a ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE_ = [coord for idx, coord in enumerate(__a ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE_ = [coord for idx, coord in enumerate(__a ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE_ = [coord for idx, coord in enumerate(__a ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE_ = [coord for idx, coord in enumerate(__a ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format SCREAMING_SNAKE_CASE_ = [] for x, y, w, h in zip(__a, __a, __a, __a ): SCREAMING_SNAKE_CASE_ = [x, y, x + w, y + h] actual_boxes.append(__a ) # finally, normalize the bounding boxes SCREAMING_SNAKE_CASE_ = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__a, __a, __a ) ) assert len(__a ) == len(__a ), "Not as many words as there are bounding boxes" return words, normalized_boxes class snake_case ( __lowercase ): UpperCAmelCase__ = ['''pixel_values'''] def __init__(self , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = PILImageResampling.BILINEAR , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = 1 / 2_55 , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = "" , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = size if size is not None else {'''height''': 2_24, '''width''': 2_24} SCREAMING_SNAKE_CASE_ = get_size_dict(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = do_resize SCREAMING_SNAKE_CASE_ = size SCREAMING_SNAKE_CASE_ = resample SCREAMING_SNAKE_CASE_ = do_rescale SCREAMING_SNAKE_CASE_ = rescale_value SCREAMING_SNAKE_CASE_ = do_normalize SCREAMING_SNAKE_CASE_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE_ = image_std if image_std is not None else IMAGENET_STANDARD_STD SCREAMING_SNAKE_CASE_ = apply_ocr SCREAMING_SNAKE_CASE_ = ocr_lang SCREAMING_SNAKE_CASE_ = tesseract_config def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = PILImageResampling.BILINEAR , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" SCREAMING_SNAKE_CASE_ = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}' ) SCREAMING_SNAKE_CASE_ = (size['''height'''], size['''width''']) return resize(SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" return rescale(SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" return normalize(SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" SCREAMING_SNAKE_CASE_ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE_ = size if size is not None else self.size SCREAMING_SNAKE_CASE_ = get_size_dict(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE_ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE_ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE_ = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE_ = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE_ = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE_ = apply_ocr if apply_ocr is not None else self.apply_ocr SCREAMING_SNAKE_CASE_ = ocr_lang if ocr_lang is not None else self.ocr_lang SCREAMING_SNAKE_CASE_ = tesseract_config if tesseract_config is not None else self.tesseract_config SCREAMING_SNAKE_CASE_ = make_list_of_images(SCREAMING_SNAKE_CASE_ ) if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''If do_normalize is True, image_mean and image_std must be specified.''' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE_ = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , '''pytesseract''' ) SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] for image in images: SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = apply_tesseract(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) words_batch.append(SCREAMING_SNAKE_CASE_ ) boxes_batch.append(SCREAMING_SNAKE_CASE_ ) if do_resize: SCREAMING_SNAKE_CASE_ = [self.resize(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE_ = [self.rescale(image=SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE_ = [self.normalize(image=SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ ) for image in images] SCREAMING_SNAKE_CASE_ = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for image in images] SCREAMING_SNAKE_CASE_ = BatchFeature(data={'''pixel_values''': images} , tensor_type=SCREAMING_SNAKE_CASE_ ) if apply_ocr: SCREAMING_SNAKE_CASE_ = words_batch SCREAMING_SNAKE_CASE_ = boxes_batch return data
626
0
'''simple docstring''' lowercase__ = [ "VerificationMode", "Version", "disable_progress_bar", "enable_progress_bar", "is_progress_bar_enabled", "experimental", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
712
'''simple docstring''' import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels lowercase__ = object() # For specifying empty leaf dict `{}` lowercase__ = object() def __UpperCamelCase ( __lowerCamelCase : Tuple , __lowerCamelCase : Tuple ) -> Union[str, Any]: '''simple docstring''' _a = tuple((re.compile(x + "$" ) for x in qs) ) for i in range(len(__lowerCamelCase ) - len(__lowerCamelCase ) + 1 ): _a = [x.match(__lowerCamelCase ) for x, y in zip(__lowerCamelCase , ks[i:] )] if matches and all(__lowerCamelCase ): return True return False def __UpperCamelCase ( __lowerCamelCase : int ) -> Union[str, Any]: '''simple docstring''' def replace(__lowerCamelCase : Tuple , __lowerCamelCase : Any ): for rule, replacement in rules: if _match(__lowerCamelCase , __lowerCamelCase ): return replacement return val return replace def __UpperCamelCase ( ) -> Tuple: '''simple docstring''' return [ # embeddings (("transformer", "wpe", "embedding"), P("mp" , __lowerCamelCase )), (("transformer", "wte", "embedding"), P("mp" , __lowerCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__lowerCamelCase , "mp" )), (("attention", "out_proj", "kernel"), P("mp" , __lowerCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__lowerCamelCase , "mp" )), (("mlp", "c_fc", "bias"), P("mp" )), (("mlp", "c_proj", "kernel"), P("mp" , __lowerCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __UpperCamelCase ( __lowerCamelCase : Dict ) -> Tuple: '''simple docstring''' _a = _get_partition_rules() _a = _replacement_rules(__lowerCamelCase ) _a = {k: _unmatched for k in flatten_dict(__lowerCamelCase )} _a = {k: replace(__lowerCamelCase , __lowerCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__lowerCamelCase ) )
276
0
'''simple docstring''' from math import ceil def lowercase_ ( _lowercase , _lowercase ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase_ : List[str] = list(range(0 , _lowercase ) ) lowerCamelCase_ : List[Any] = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check lowerCamelCase_ : Optional[Any] = [] for i in device_map_blocks: if device_map_blocks.count(_lowercase ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(_lowercase ) # Missing blocks lowerCamelCase_ : Dict = [i for i in blocks if i not in device_map_blocks] lowerCamelCase_ : List[str] = [i for i in device_map_blocks if i not in blocks] if len(_lowercase ) != 0: raise ValueError( '''Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.''' ''' These attention blocks were specified more than once: ''' + str(_lowercase ) ) if len(_lowercase ) != 0: raise ValueError( '''There are attention blocks for this model that are not specified in the device_map. Add these attention ''' '''blocks to a device on the device_map: ''' + str(_lowercase ) ) if len(_lowercase ) != 0: raise ValueError( '''The device_map contains more attention blocks than this model has. Remove these from the device_map:''' + str(_lowercase ) ) def lowercase_ ( _lowercase , _lowercase ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase_ : Optional[int] = list(range(_lowercase ) ) lowerCamelCase_ : str = int(ceil(n_layers / len(_lowercase ) ) ) lowerCamelCase_ : str = [layers[i : i + n_blocks] for i in range(0 , _lowercase , _lowercase )] return dict(zip(_lowercase , _lowercase ) )
422
'''simple docstring''' import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowercase_ ( _lowercase , _lowercase , _lowercase ) -> Tuple: '''simple docstring''' if openai_config_file == "": lowerCamelCase_ : str = OpenAIGPTConfig() else: lowerCamelCase_ : List[str] = OpenAIGPTConfig.from_json_file(_lowercase ) lowerCamelCase_ : Any = OpenAIGPTModel(_lowercase ) # Load weights from numpy load_tf_weights_in_openai_gpt(_lowercase , _lowercase , _lowercase ) # Save pytorch-model lowerCamelCase_ : Union[str, Any] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME lowerCamelCase_ : Union[str, Any] = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , _lowercase ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(_lowercase , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowercase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--openai_checkpoint_folder_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--openai_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained OpenAI model. \n''' '''This specifies the model architecture.''' ), ) __lowercase : List[Any] = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
422
1
'''simple docstring''' from __future__ import annotations __lowerCamelCase : Dict = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,)-> tuple[list[list[int]], list[list[int]]]: """simple docstring""" snake_case_ : List[Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the reference grid snake_case_ : Optional[Any] = 1 snake_case_ : Union[str, Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the action grid snake_case_ : List[Any] = init[0] snake_case_ : List[Any] = init[1] snake_case_ : Any = 0 snake_case_ : Tuple = g + heuristic[x][y] # cost from starting cell to destination cell snake_case_ : Tuple = [[f, g, x, y]] snake_case_ : List[Any] = False # flag that is set when search is complete snake_case_ : Tuple = False # flag set if we can't find expand while not found and not resign: if len(__magic_name__ ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() snake_case_ : Tuple = cell.pop() snake_case_ : List[str] = next_cell[2] snake_case_ : int = next_cell[3] snake_case_ : Optional[int] = next_cell[1] if x == goal[0] and y == goal[1]: snake_case_ : Tuple = True else: for i in range(len(__magic_name__ ) ): # to try out different valid actions snake_case_ : Optional[Any] = x + DIRECTIONS[i][0] snake_case_ : List[str] = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__magic_name__ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: snake_case_ : Dict = g + cost snake_case_ : Any = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) snake_case_ : Union[str, Any] = 1 snake_case_ : Optional[Any] = i snake_case_ : Optional[Any] = [] snake_case_ : Any = goal[0] snake_case_ : Tuple = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: snake_case_ : Any = x - DIRECTIONS[action[x][y]][0] snake_case_ : Tuple = y - DIRECTIONS[action[x][y]][1] snake_case_ : Tuple = xa snake_case_ : Tuple = ya invpath.append([x, y] ) snake_case_ : Any = [] for i in range(len(__magic_name__ ) ): path.append(invpath[len(__magic_name__ ) - 1 - i] ) return path, action if __name__ == "__main__": __lowerCamelCase : Optional[int] = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] __lowerCamelCase : Any = [0, 0] # all coordinates are given in format [y,x] __lowerCamelCase : Optional[Any] = [len(grid) - 1, len(grid[0]) - 1] __lowerCamelCase : Any = 1 # the cost map which pushes the path closer to the goal __lowerCamelCase : int = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): __lowerCamelCase : Union[str, Any] = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map __lowerCamelCase : Dict = 99 __lowerCamelCase , __lowerCamelCase : Dict = search(grid, init, goal, cost, heuristic) print('''ACTION MAP''') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
656
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } snake_case_ : int = Dataset.from_dict(__magic_name__ ) return dataset class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = get_dataset() snake_case_ : Optional[int] = make_duplicate_clusters(lowerCAmelCase__ , 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = get_dataset() snake_case_, snake_case_ : List[Any] = deduplicate_dataset(lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 2 ) print(lowerCAmelCase__ ) self.assertEqual(duplicate_clusters[0][0]["copies"] , 2 ) self.assertEqual(duplicate_clusters[0][0]["is_extreme"] , lowerCAmelCase__ )
656
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _a: str = logging.get_logger(__name__) _a: Dict = { """asapp/sew-tiny-100k""": """https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json""", # See all SEW models at https://huggingface.co/models?filter=sew } class __UpperCamelCase ( lowercase ): SCREAMING_SNAKE_CASE__ = 'sew' def __init__( self : Any , lowerCAmelCase : int=32 , lowerCAmelCase : Tuple=768 , lowerCAmelCase : Optional[int]=12 , lowerCAmelCase : int=12 , lowerCAmelCase : Union[str, Any]=3_072 , lowerCAmelCase : List[str]=2 , lowerCAmelCase : Tuple="gelu" , lowerCAmelCase : int=0.1 , lowerCAmelCase : List[Any]=0.1 , lowerCAmelCase : int=0.1 , lowerCAmelCase : Union[str, Any]=0.0 , lowerCAmelCase : Optional[int]=0.1 , lowerCAmelCase : Optional[int]=0.1 , lowerCAmelCase : List[Any]=0.02 , lowerCAmelCase : Dict=1e-5 , lowerCAmelCase : Optional[int]="group" , lowerCAmelCase : Dict="gelu" , lowerCAmelCase : List[str]=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowerCAmelCase : str=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowerCAmelCase : str=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowerCAmelCase : Any=False , lowerCAmelCase : Optional[int]=128 , lowerCAmelCase : Optional[int]=16 , lowerCAmelCase : Union[str, Any]=True , lowerCAmelCase : Union[str, Any]=0.05 , lowerCAmelCase : str=10 , lowerCAmelCase : Any=2 , lowerCAmelCase : Tuple=0.0 , lowerCAmelCase : int=10 , lowerCAmelCase : List[str]=0 , lowerCAmelCase : int="mean" , lowerCAmelCase : str=False , lowerCAmelCase : Optional[Any]=False , lowerCAmelCase : str=256 , lowerCAmelCase : int=0 , lowerCAmelCase : str=1 , lowerCAmelCase : Optional[int]=2 , **lowerCAmelCase : Optional[Any] , ): '''simple docstring''' super().__init__(**lowerCAmelCase , pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase ) UpperCAmelCase_ = hidden_size UpperCAmelCase_ = feat_extract_norm UpperCAmelCase_ = feat_extract_activation UpperCAmelCase_ = list(lowerCAmelCase ) UpperCAmelCase_ = list(lowerCAmelCase ) UpperCAmelCase_ = list(lowerCAmelCase ) UpperCAmelCase_ = conv_bias UpperCAmelCase_ = num_conv_pos_embeddings UpperCAmelCase_ = num_conv_pos_embedding_groups UpperCAmelCase_ = len(self.conv_dim ) UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = squeeze_factor UpperCAmelCase_ = hidden_act UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_dropout UpperCAmelCase_ = attention_dropout UpperCAmelCase_ = activation_dropout UpperCAmelCase_ = feat_proj_dropout UpperCAmelCase_ = final_dropout UpperCAmelCase_ = layerdrop UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = initializer_range UpperCAmelCase_ = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase_ = apply_spec_augment UpperCAmelCase_ = mask_time_prob UpperCAmelCase_ = mask_time_length UpperCAmelCase_ = mask_time_min_masks UpperCAmelCase_ = mask_feature_prob UpperCAmelCase_ = mask_feature_length UpperCAmelCase_ = mask_feature_min_masks # ctc loss UpperCAmelCase_ = ctc_loss_reduction UpperCAmelCase_ = ctc_zero_infinity # sequence classification UpperCAmelCase_ = use_weighted_layer_sum UpperCAmelCase_ = classifier_proj_size @property def __A ( self : str ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
162
from __future__ import annotations import pandas as pd def __lowerCAmelCase ( A , A , A ): UpperCAmelCase_ = [0] * no_of_processes UpperCAmelCase_ = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(A ): UpperCAmelCase_ = burst_time[i] UpperCAmelCase_ = 0 UpperCAmelCase_ = 0 UpperCAmelCase_ = 999999999 UpperCAmelCase_ = 0 UpperCAmelCase_ = False # Process until all processes are completed while complete != no_of_processes: for j in range(A ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: UpperCAmelCase_ = remaining_time[j] UpperCAmelCase_ = j UpperCAmelCase_ = True if not check: increment_time += 1 continue remaining_time[short] -= 1 UpperCAmelCase_ = remaining_time[short] if minm == 0: UpperCAmelCase_ = 999999999 if remaining_time[short] == 0: complete += 1 UpperCAmelCase_ = False # Find finish time of current process UpperCAmelCase_ = increment_time + 1 # Calculate waiting time UpperCAmelCase_ = finish_time - arrival_time[short] UpperCAmelCase_ = finar - burst_time[short] if waiting_time[short] < 0: UpperCAmelCase_ = 0 # Increment time increment_time += 1 return waiting_time def __lowerCAmelCase ( A , A , A ): UpperCAmelCase_ = [0] * no_of_processes for i in range(A ): UpperCAmelCase_ = burst_time[i] + waiting_time[i] return turn_around_time def __lowerCAmelCase ( A , A , A ): UpperCAmelCase_ = 0 UpperCAmelCase_ = 0 for i in range(A ): UpperCAmelCase_ = total_waiting_time + waiting_time[i] UpperCAmelCase_ = total_turn_around_time + turn_around_time[i] print(F"Average waiting time = {total_waiting_time / no_of_processes:.5f}" ) print("Average turn around time =" , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") _a: str = int(input()) _a: List[str] = [0] * no_of_processes _a: Dict = [0] * no_of_processes _a: Optional[int] = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) _a , _a: List[str] = map(int, input().split()) _a: List[str] = calculate_waitingtime(arrival_time, burst_time, no_of_processes) _a: Optional[int] = burst_time _a: List[Any] = no_of_processes _a: List[Any] = waiting_time _a: List[Any] = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) _a: Any = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
162
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A__ = { """configuration_blenderbot_small""": [ """BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlenderbotSmallConfig""", """BlenderbotSmallOnnxConfig""", ], """tokenization_blenderbot_small""": ["""BlenderbotSmallTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = ["""BlenderbotSmallTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ """BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlenderbotSmallForCausalLM""", """BlenderbotSmallForConditionalGeneration""", """BlenderbotSmallModel""", """BlenderbotSmallPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ """TFBlenderbotSmallForConditionalGeneration""", """TFBlenderbotSmallModel""", """TFBlenderbotSmallPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ """FlaxBlenderbotSmallForConditionalGeneration""", """FlaxBlenderbotSmallModel""", """FlaxBlenderbotSmallPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys A__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
49
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A__ = { """configuration_layoutlmv2""": ["""LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LayoutLMv2Config"""], """processing_layoutlmv2""": ["""LayoutLMv2Processor"""], """tokenization_layoutlmv2""": ["""LayoutLMv2Tokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = ["""LayoutLMv2TokenizerFast"""] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = ["""LayoutLMv2FeatureExtractor"""] A__ = ["""LayoutLMv2ImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ """LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """LayoutLMv2ForQuestionAnswering""", """LayoutLMv2ForSequenceClassification""", """LayoutLMv2ForTokenClassification""", """LayoutLMv2Layer""", """LayoutLMv2Model""", """LayoutLMv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys A__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
49
1
'''simple docstring''' from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def A__ ( A_ ) -> int: _lowercase = prime_factors(A_ ) if is_square_free(A_ ): return -1 if len(A_ ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
497
'''simple docstring''' import os import sys __magic_name__ : str = os.path.join(os.path.dirname(__file__), '''src''') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __magic_name__ : List[Any] = [ '''torch''', '''numpy''', '''tokenizers''', '''filelock''', '''requests''', '''tqdm''', '''regex''', '''sentencepiece''', '''sacremoses''', '''importlib_metadata''', '''huggingface_hub''', ] @add_start_docstrings(AutoConfig.__doc__ ) def A__ ( *A_ , **A_ ) -> List[str]: return AutoConfig.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def A__ ( *A_ , **A_ ) -> str: return AutoTokenizer.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoModel.__doc__ ) def A__ ( *A_ , **A_ ) -> Dict: return AutoModel.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def A__ ( *A_ , **A_ ) -> int: return AutoModelForCausalLM.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def A__ ( *A_ , **A_ ) -> int: return AutoModelForMaskedLM.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def A__ ( *A_ , **A_ ) -> Any: return AutoModelForSequenceClassification.from_pretrained(*A_ , **A_ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def A__ ( *A_ , **A_ ) -> str: return AutoModelForQuestionAnswering.from_pretrained(*A_ , **A_ )
497
1
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input __A : List[str] = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def __UpperCamelCase ( ) ->List[str]: """simple docstring""" lowerCamelCase_ =_ask_options( """In which compute environment are you running?""" , ["""This machine""", """AWS (Amazon SageMaker)"""] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: lowerCamelCase_ =get_sagemaker_input() else: lowerCamelCase_ =get_cluster_input() return config def __UpperCamelCase ( _A : List[str]=None ) ->str: """simple docstring""" if subparsers is not None: lowerCamelCase_ =subparsers.add_parser("""config""" , description=_A ) else: lowerCamelCase_ =argparse.ArgumentParser("""Accelerate config command""" , description=_A ) parser.add_argument( """--config_file""" , default=_A , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ) , ) if subparsers is not None: parser.set_defaults(func=_A ) return parser def __UpperCamelCase ( _A : Union[str, Any] ) ->Optional[int]: """simple docstring""" lowerCamelCase_ =get_user_input() if args.config_file is not None: lowerCamelCase_ =args.config_file else: if not os.path.isdir(_A ): os.makedirs(_A ) lowerCamelCase_ =default_yaml_config_file if config_file.endswith(""".json""" ): config.to_json_file(_A ) else: config.to_yaml_file(_A ) print(f'accelerate configuration saved at {config_file}' ) def __UpperCamelCase ( ) ->Dict: """simple docstring""" lowerCamelCase_ =config_command_parser() lowerCamelCase_ =parser.parse_args() config_command(_A ) if __name__ == "__main__": main()
720
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __A : Tuple = {'configuration_reformer': ['REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ReformerConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = ['ReformerTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = ['ReformerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = [ 'REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ReformerAttention', 'ReformerForMaskedLM', 'ReformerForQuestionAnswering', 'ReformerForSequenceClassification', 'ReformerLayer', 'ReformerModel', 'ReformerModelWithLMHead', 'ReformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_reformer import REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ReformerConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer import ReformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer_fast import ReformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_reformer import ( REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ReformerAttention, ReformerForMaskedLM, ReformerForQuestionAnswering, ReformerForSequenceClassification, ReformerLayer, ReformerModel, ReformerModelWithLMHead, ReformerPreTrainedModel, ) else: import sys __A : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
75
0
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class lowerCamelCase_ ( unittest.TestCase ): def lowercase_ ( self : Tuple , _A : Dict , _A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = jnp.ones((batch_size, length) ) / length return scores def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : Optional[Any] = 20 UpperCAmelCase__ : Union[str, Any] = self._get_uniform_logits(batch_size=2 , length=SCREAMING_SNAKE_CASE_ ) # tweak scores to not be uniform anymore UpperCAmelCase__ : Optional[Any] = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch UpperCAmelCase__ : Optional[int] = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax UpperCAmelCase__ : Any = jax.nn.softmax(SCREAMING_SNAKE_CASE_ , axis=-1 ) UpperCAmelCase__ : str = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCAmelCase__ : List[Any] = FlaxTemperatureLogitsWarper(temperature=1.3 ) UpperCAmelCase__ : Any = jax.nn.softmax(temp_dist_warper_sharper(SCREAMING_SNAKE_CASE_ , scores.copy() , cur_len=SCREAMING_SNAKE_CASE_ ) , axis=-1 ) UpperCAmelCase__ : Any = jax.nn.softmax(temp_dist_warper_smoother(SCREAMING_SNAKE_CASE_ , scores.copy() , cur_len=SCREAMING_SNAKE_CASE_ ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1e-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1e-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def lowercase_ ( self : str ): '''simple docstring''' UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = 10 UpperCAmelCase__ : Optional[int] = 2 # create ramp distribution UpperCAmelCase__ : int = np.broadcast_to(np.arange(SCREAMING_SNAKE_CASE_ )[None, :] , (batch_size, vocab_size) ).copy() UpperCAmelCase__ : List[Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size UpperCAmelCase__ : List[Any] = FlaxTopKLogitsWarper(3 ) UpperCAmelCase__ : Any = top_k_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case UpperCAmelCase__ : int = 5 UpperCAmelCase__ : Union[str, Any] = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) UpperCAmelCase__ : int = np.broadcast_to(np.arange(SCREAMING_SNAKE_CASE_ )[None, :] , (batch_size, length) ).copy() UpperCAmelCase__ : List[Any] = top_k_warp_safety_check(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : int = None UpperCAmelCase__ : Dict = 10 UpperCAmelCase__ : Optional[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) UpperCAmelCase__ : Dict = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.1_5, 0.3, 0.3, 0.2_5]] ) ) UpperCAmelCase__ : List[Any] = FlaxTopPLogitsWarper(0.8 ) UpperCAmelCase__ : Union[str, Any] = np.exp(top_p_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 UpperCAmelCase__ : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.2_5]] ) self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) # check edge cases with negative and extreme logits UpperCAmelCase__ : Tuple = np.broadcast_to(np.arange(SCREAMING_SNAKE_CASE_ )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme UpperCAmelCase__ : Union[str, Any] = ramp_logits[1] * 1_0_0.0 # make sure at least 2 tokens are kept UpperCAmelCase__ : str = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) UpperCAmelCase__ : str = top_p_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = 20 UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : List[Any] = 0 UpperCAmelCase__ : int = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=SCREAMING_SNAKE_CASE_ ) # check that min length is applied at length 5 UpperCAmelCase__ : Union[str, Any] = ids_tensor((batch_size, 20) , vocab_size=20 ) UpperCAmelCase__ : Optional[int] = 5 UpperCAmelCase__ : int = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : int = min_dist_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 UpperCAmelCase__ : int = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Optional[int] = 15 UpperCAmelCase__ : Optional[int] = min_dist_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertFalse(jnp.isinf(SCREAMING_SNAKE_CASE_ ).any() ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Dict = 20 UpperCAmelCase__ : Any = 4 UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : str = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=SCREAMING_SNAKE_CASE_ ) # check that all scores are -inf except the bos_token_id score UpperCAmelCase__ : int = ids_tensor((batch_size, 1) , vocab_size=20 ) UpperCAmelCase__ : Optional[int] = 1 UpperCAmelCase__ : Optional[int] = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : List[str] = logits_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 UpperCAmelCase__ : Union[str, Any] = 3 UpperCAmelCase__ : Dict = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Union[str, Any] = logits_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertFalse(jnp.isinf(SCREAMING_SNAKE_CASE_ ).any() ) def lowercase_ ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = 20 UpperCAmelCase__ : Optional[Any] = 4 UpperCAmelCase__ : List[str] = 0 UpperCAmelCase__ : int = 5 UpperCAmelCase__ : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ ) # check that all scores are -inf except the eos_token_id when max_length is reached UpperCAmelCase__ : str = ids_tensor((batch_size, 4) , vocab_size=20 ) UpperCAmelCase__ : List[Any] = 4 UpperCAmelCase__ : List[Any] = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Union[str, Any] = logits_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached UpperCAmelCase__ : List[Any] = 3 UpperCAmelCase__ : List[str] = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Tuple = logits_processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) self.assertFalse(jnp.isinf(SCREAMING_SNAKE_CASE_ ).any() ) def lowercase_ ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = 4 UpperCAmelCase__ : Tuple = 10 UpperCAmelCase__ : Optional[int] = 15 UpperCAmelCase__ : Dict = 2 UpperCAmelCase__ : List[Any] = 1 UpperCAmelCase__ : int = 15 # dummy input_ids and scores UpperCAmelCase__ : Dict = ids_tensor((batch_size, sequence_length) , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Tuple = input_ids.copy() UpperCAmelCase__ : int = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : List[Any] = scores.copy() # instantiate all dist processors UpperCAmelCase__ : Any = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCAmelCase__ : Dict = FlaxTopKLogitsWarper(3 ) UpperCAmelCase__ : List[str] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors UpperCAmelCase__ : List[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Union[str, Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Dict = 10 # no processor list UpperCAmelCase__ : Tuple = temp_dist_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Tuple = top_k_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : str = top_p_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : int = min_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Tuple = bos_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Dict = eos_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) # with processor list UpperCAmelCase__ : str = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) UpperCAmelCase__ : Union[str, Any] = processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) # scores should be equal self.assertTrue(jnp.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = 4 UpperCAmelCase__ : str = 10 UpperCAmelCase__ : List[str] = 15 UpperCAmelCase__ : Any = 2 UpperCAmelCase__ : str = 1 UpperCAmelCase__ : Optional[int] = 15 # dummy input_ids and scores UpperCAmelCase__ : List[str] = ids_tensor((batch_size, sequence_length) , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Any = input_ids.copy() UpperCAmelCase__ : Union[str, Any] = self._get_uniform_logits(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Any = scores.copy() # instantiate all dist processors UpperCAmelCase__ : Any = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCAmelCase__ : Union[str, Any] = FlaxTopKLogitsWarper(3 ) UpperCAmelCase__ : Tuple = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors UpperCAmelCase__ : List[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Dict = 10 # no processor list def run_no_processor_list(_A : List[Any] , _A : List[str] , _A : Union[str, Any] ): UpperCAmelCase__ : Dict = temp_dist_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : str = top_k_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : int = top_p_warp(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Dict = min_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : List[str] = bos_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Tuple = eos_dist_proc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) return scores # with processor list def run_processor_list(_A : Dict , _A : List[Any] , _A : Dict ): UpperCAmelCase__ : int = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) UpperCAmelCase__ : List[Any] = processor(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , cur_len=SCREAMING_SNAKE_CASE_ ) return scores UpperCAmelCase__ : List[Any] = jax.jit(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : List[Any] = jax.jit(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Union[str, Any] = jitted_run_no_processor_list(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase__ : Union[str, Any] = jitted_run_processor_list(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # scores should be equal self.assertTrue(jnp.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
75
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __magic_name__ =logging.get_logger(__name__) class _A ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ : Any =["pixel_values"] def __init__(self , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = 1 / 255 , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> None: '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = size if size is not None else {'''height''': 256, '''width''': 256} UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' ) UpperCamelCase__ = do_resize UpperCamelCase__ = size UpperCamelCase__ = resample UpperCamelCase__ = do_center_crop UpperCamelCase__ = crop_size UpperCamelCase__ = do_rescale UpperCamelCase__ = rescale_factor UpperCamelCase__ = do_normalize UpperCamelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCamelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must have keys 'height' and 'width'. Got {size.keys()}" ) return resize( SCREAMING_SNAKE_CASE_ , size=(size['''height'''], size['''width''']) , resample=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(SCREAMING_SNAKE_CASE_ , size=(size['''height'''], size['''width''']) , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> List[str]: '''simple docstring''' return rescale(SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' return normalize(SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE_ , ) -> PIL.Image.Image: '''simple docstring''' UpperCamelCase__ = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ = resample if resample is not None else self.resample UpperCamelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase__ = image_mean if image_mean is not None else self.image_mean UpperCamelCase__ = image_std if image_std is not None else self.image_std UpperCamelCase__ = size if size is not None else self.size UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = crop_size if crop_size is not None else self.crop_size UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' ) UpperCamelCase__ = make_list_of_images(SCREAMING_SNAKE_CASE_ ) if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. UpperCamelCase__ = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] if do_resize: UpperCamelCase__ = [self.resize(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_center_crop: UpperCamelCase__ = [self.center_crop(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: UpperCamelCase__ = [self.rescale(image=SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: UpperCamelCase__ = [self.normalize(image=SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase__ = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase__ = {'''pixel_values''': images} return BatchFeature(data=SCREAMING_SNAKE_CASE_ , tensor_type=SCREAMING_SNAKE_CASE_ )
415
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer _lowercase : List[str] = logging.get_logger(__name__) _lowercase : Union[str, Any] = {'vocab_file': 'vocab.txt'} _lowercase : str = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } _lowercase : List[str] = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } _lowercase : Tuple = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class UpperCamelCase__( _A ): __magic_name__ : Optional[int] = VOCAB_FILES_NAMES __magic_name__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __magic_name__ : Optional[int] = PRETRAINED_INIT_CONFIGURATION __magic_name__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ : List[Any] = ConvBertTokenizer def __init__( self : str , lowerCAmelCase : Optional[int]=None , lowerCAmelCase : List[Any]=None , lowerCAmelCase : Dict=True , lowerCAmelCase : Optional[Any]="[UNK]" , lowerCAmelCase : Optional[Any]="[SEP]" , lowerCAmelCase : Tuple="[PAD]" , lowerCAmelCase : Any="[CLS]" , lowerCAmelCase : int="[MASK]" , lowerCAmelCase : Any=True , lowerCAmelCase : Optional[Any]=None , **lowerCAmelCase : Dict , )-> List[Any]: """simple docstring""" super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenize_chinese_chars=UpperCamelCase__ , strip_accents=UpperCamelCase__ , **UpperCamelCase__ , ) UpperCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCamelCase__ ) != tokenize_chinese_chars ): UpperCAmelCase = getattr(UpperCamelCase__ , normalizer_state.pop('''type''' ) ) UpperCAmelCase = do_lower_case UpperCAmelCase = strip_accents UpperCAmelCase = tokenize_chinese_chars UpperCAmelCase = normalizer_class(**UpperCamelCase__ ) UpperCAmelCase = do_lower_case def a__( self : int , lowerCAmelCase : int , lowerCAmelCase : str=None )-> Optional[int]: """simple docstring""" UpperCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def a__( self : List[Any] , lowerCAmelCase : List[int] , lowerCAmelCase : Optional[List[int]] = None )-> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a__( self : Tuple , lowerCAmelCase : str , lowerCAmelCase : Optional[str] = None )-> Tuple[str]: """simple docstring""" UpperCAmelCase = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ )
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase : Any = { """configuration_maskformer""": ["""MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MaskFormerConfig"""], """configuration_maskformer_swin""": ["""MaskFormerSwinConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Any = ["""MaskFormerFeatureExtractor"""] _lowercase : Dict = ["""MaskFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Any = [ """MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """MaskFormerForInstanceSegmentation""", """MaskFormerModel""", """MaskFormerPreTrainedModel""", ] _lowercase : List[Any] = [ """MaskFormerSwinBackbone""", """MaskFormerSwinModel""", """MaskFormerSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys _lowercase : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
50
0
from math import pi, sqrt, tan def a__ ( lowercase__ ): '''simple docstring''' if side_length < 0: raise ValueError("surface_area_cube() only accepts non-negative values" ) return 6 * side_length**2 def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError("surface_area_cuboid() only accepts non-negative values" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("surface_area_sphere() only accepts non-negative values" ) return 4 * pi * radius**2 def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("surface_area_hemisphere() only accepts non-negative values" ) return 3 * pi * radius**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("surface_area_cone() only accepts non-negative values" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( "surface_area_conical_frustum() only accepts non-negative values" ) UpperCAmelCase_ =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("surface_area_cylinder() only accepts non-negative values" ) return 2 * pi * radius * (height + radius) def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError("surface_area_torus() only accepts non-negative values" ) if torus_radius < tube_radius: raise ValueError( "surface_area_torus() does not support spindle or self intersecting tori" ) return 4 * pow(lowercase__ , 2 ) * torus_radius * tube_radius def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if length < 0 or width < 0: raise ValueError("area_rectangle() only accepts non-negative values" ) return length * width def a__ ( lowercase__ ): '''simple docstring''' if side_length < 0: raise ValueError("area_square() only accepts non-negative values" ) return side_length**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("area_triangle() only accepts non-negative values" ) return (base * height) / 2 def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("area_triangle_three_sides() only accepts non-negative values" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("Given three sides do not form a triangle" ) UpperCAmelCase_ =(sidea + sidea + sidea) / 2 UpperCAmelCase_ =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("area_parallelogram() only accepts non-negative values" ) return base * height def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError("area_trapezium() only accepts non-negative values" ) return 1 / 2 * (basea + basea) * height def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("area_circle() only accepts non-negative values" ) return pi * radius**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError("area_ellipse() only accepts non-negative values" ) return pi * radius_x * radius_y def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError("area_rhombus() only accepts non-negative values" ) return 1 / 2 * diagonal_a * diagonal_a def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ) or sides < 3: raise ValueError( "area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides" ) elif length < 0: raise ValueError( "area_reg_polygon() only accepts non-negative values as \ length of a side" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("""[DEMO] Areas of various geometric shapes: \n""") print(f"""Rectangle: {area_rectangle(10, 20) = }""") print(f"""Square: {area_square(10) = }""") print(f"""Triangle: {area_triangle(10, 10) = }""") print(f"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(f"""Parallelogram: {area_parallelogram(10, 20) = }""") print(f"""Rhombus: {area_rhombus(10, 20) = }""") print(f"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(f"""Circle: {area_circle(20) = }""") print(f"""Ellipse: {area_ellipse(10, 20) = }""") print("""\nSurface Areas of various geometric shapes: \n""") print(f"""Cube: {surface_area_cube(20) = }""") print(f"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(f"""Sphere: {surface_area_sphere(20) = }""") print(f"""Hemisphere: {surface_area_hemisphere(20) = }""") print(f"""Cone: {surface_area_cone(10, 20) = }""") print(f"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(f"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(f"""Torus: {surface_area_torus(20, 10) = }""") print(f"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(f"""Square: {area_reg_polygon(4, 10) = }""") print(f"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
54
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from math import pow, sqrt def lowerCamelCase__ ( *a : float ) -> bool: """simple docstring""" a__ :Optional[int] = len(a ) > 0 and all(value > 0.0 for value in values ) return result def lowerCamelCase__ ( a : float , a : float ) -> float | ValueError: """simple docstring""" return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(a , a ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def lowerCamelCase__ ( a : float , a : float , a : float ) -> float | ValueError: """simple docstring""" return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(a , a , a ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def lowerCamelCase__ ( a : float , a : float , a : float ) -> float | ValueError: """simple docstring""" return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(a , a , a ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def lowerCamelCase__ ( a : float , a : float , a : float ) -> float | ValueError: """simple docstring""" return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(a , a , a ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def lowerCamelCase__ ( a : float , a : float , a : float ) -> float | ValueError: """simple docstring""" return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(a , a , a ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
373
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device snake_case__ = False class lowerCAmelCase_ ( unittest.TestCase): pass @nightly @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase): def _snake_case ( self : Tuple ) ->Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self : Dict ) ->Any: """simple docstring""" a__ :Union[str, Any] = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion" ) # remove text_unet pipe.remove_unused_weights() pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) a__ :List[Any] = "A painting of a squirrel eating a burger " a__ :Optional[Any] = torch.manual_seed(0 ) a__ :List[Any] = pipe( prompt=__A , generator=__A , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__A ) a__ :List[Any] = VersatileDiffusionTextToImagePipeline.from_pretrained(__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) a__ :Optional[int] = generator.manual_seed(0 ) a__ :List[Any] = pipe( prompt=__A , generator=__A , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _snake_case ( self : Optional[Any] ) ->List[Any]: """simple docstring""" a__ :Tuple = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) a__ :Tuple = "A painting of a squirrel eating a burger " a__ :Tuple = torch.manual_seed(0 ) a__ :Optional[Any] = pipe( prompt=__A , generator=__A , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images a__ :Tuple = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) a__ :Tuple = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
373
1
import string def snake_case_ (__A : Optional[int] ) -> Union[str, Any]: for key in range(len(string.ascii_uppercase ) ): __lowerCAmelCase : Dict = "" for symbol in message: if symbol in string.ascii_uppercase: __lowerCAmelCase : Union[str, Any] = string.ascii_uppercase.find(__A ) __lowerCAmelCase : Dict = num - key if num < 0: __lowerCAmelCase : List[str] = num + len(string.ascii_uppercase ) __lowerCAmelCase : int = translated + string.ascii_uppercase[num] else: __lowerCAmelCase : int = translated + symbol print(f'''Decryption using Key #{key}: {translated}''' ) def snake_case_ () -> Union[str, Any]: __lowerCAmelCase : Any = input("""Encrypted message: """ ) __lowerCAmelCase : int = message.upper() decrypt(__A ) if __name__ == "__main__": import doctest doctest.testmod() main()
651
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def A_ ( snake_case , snake_case , snake_case , snake_case , snake_case ): # Load configuration defined in the metadata file with open(snake_case ) as metadata_file: SCREAMING_SNAKE_CASE:str = json.load(snake_case ) SCREAMING_SNAKE_CASE:List[str] = LukeConfig(use_entity_aware_attention=snake_case , **metadata["model_config"] ) # Load in the weights from the checkpoint_path SCREAMING_SNAKE_CASE:Tuple = torch.load(snake_case , map_location="cpu" ) # Load the entity vocab file SCREAMING_SNAKE_CASE:Dict = load_entity_vocab(snake_case ) SCREAMING_SNAKE_CASE:Optional[int] = RobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks SCREAMING_SNAKE_CASE:Dict = AddedToken("<ent>" , lstrip=snake_case , rstrip=snake_case ) SCREAMING_SNAKE_CASE:List[Any] = AddedToken("<ent2>" , lstrip=snake_case , rstrip=snake_case ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(snake_case ) with open(os.path.join(snake_case , LukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(snake_case , snake_case ) SCREAMING_SNAKE_CASE:str = LukeTokenizer.from_pretrained(snake_case ) # Initialize the embeddings of the special tokens SCREAMING_SNAKE_CASE:Optional[Any] = state_dict["embeddings.word_embeddings.weight"] SCREAMING_SNAKE_CASE:Union[str, Any] = word_emb[tokenizer.convert_tokens_to_ids(["@"] )[0]].unsqueeze(0 ) SCREAMING_SNAKE_CASE:Union[str, Any] = word_emb[tokenizer.convert_tokens_to_ids(["#"] )[0]].unsqueeze(0 ) SCREAMING_SNAKE_CASE:Tuple = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: SCREAMING_SNAKE_CASE:Union[str, Any] = F'''encoder.layer.{layer_index}.attention.self.''' SCREAMING_SNAKE_CASE:List[Any] = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE:List[Any] = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE:List[str] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks SCREAMING_SNAKE_CASE:str = state_dict["entity_embeddings.entity_embeddings.weight"] SCREAMING_SNAKE_CASE:int = entity_emb[entity_vocab["[MASK]"]] SCREAMING_SNAKE_CASE:str = LukeModel(config=snake_case ).eval() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE:int = model.load_state_dict(snake_case , strict=snake_case ) if not (len(snake_case ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {", ".join(snake_case )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith("entity_predictions" ) or key.startswith("lm_head" ) for key in unexpected_keys )): raise ValueError( "Unexpected keys" F''' {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}''' ) # Check outputs SCREAMING_SNAKE_CASE:Optional[Any] = LukeTokenizer.from_pretrained(snake_case , task="entity_classification" ) SCREAMING_SNAKE_CASE:Tuple = ( "Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the" " new world number one avoid a humiliating second- round exit at Wimbledon ." ) SCREAMING_SNAKE_CASE:List[str] = (39, 42) SCREAMING_SNAKE_CASE:int = tokenizer(snake_case , entity_spans=[span] , add_prefix_space=snake_case , return_tensors="pt" ) SCREAMING_SNAKE_CASE:Any = model(**snake_case ) # Verify word hidden states if model_size == "large": SCREAMING_SNAKE_CASE:List[str] = torch.Size((1, 42, 1024) ) SCREAMING_SNAKE_CASE:Optional[int] = torch.tensor( [[0.0133, 0.0865, 0.0095], [0.3093, -0.2576, -0.7418], [-0.1720, -0.2117, -0.2869]] ) else: # base SCREAMING_SNAKE_CASE:List[str] = torch.Size((1, 42, 768) ) SCREAMING_SNAKE_CASE:int = torch.tensor([[0.0037, 0.1368, -0.0091], [0.1099, 0.3329, -0.1095], [0.0765, 0.5335, 0.1179]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": SCREAMING_SNAKE_CASE:Union[str, Any] = torch.Size((1, 1, 1024) ) SCREAMING_SNAKE_CASE:Optional[int] = torch.tensor([[0.0466, -0.0106, -0.0179]] ) else: # base SCREAMING_SNAKE_CASE:List[str] = torch.Size((1, 1, 768) ) SCREAMING_SNAKE_CASE:Any = torch.tensor([[0.1457, 0.1044, 0.0174]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , snake_case , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(snake_case ) ) model.save_pretrained(snake_case ) def A_ ( snake_case ): SCREAMING_SNAKE_CASE:Any = {} with open(snake_case , "r" , encoding="utf-8" ) as f: for index, line in enumerate(snake_case ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE:Optional[Any] = line.rstrip().split("\t" ) SCREAMING_SNAKE_CASE:str = index return entity_vocab if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument("--checkpoint_path", type=str, help="Path to a pytorch_model.bin file.") parser.add_argument( "--metadata_path", default=None, type=str, help="Path to a metadata.json file, defining the configuration." ) parser.add_argument( "--entity_vocab_path", default=None, type=str, help="Path to an entity_vocab.tsv file, containing the entity vocabulary.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to where to dump the output PyTorch model." ) parser.add_argument( "--model_size", default="base", type=str, choices=["base", "large"], help="Size of the model to be converted." ) A_ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
143
0
"""simple docstring""" def A_ ( __UpperCamelCase : int , __UpperCamelCase : int ): return int((input_a, input_a).count(1 ) != 0 ) def A_ ( ): assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
396
"""simple docstring""" import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging __lowerCAmelCase = logging.get_logger(__name__) def A_ ( ): # Get the sagemaker specific mp parameters from smp_options variable. lowercase = os.getenv('''SM_HP_MP_PARAMETERS''' , '''{}''' ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. lowercase = json.loads(__UpperCamelCase ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. lowercase = os.getenv('''SM_FRAMEWORK_PARAMS''' , '''{}''' ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". lowercase = json.loads(__UpperCamelCase ) if not mpi_options.get('''sagemaker_mpi_enabled''' , __UpperCamelCase ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec('''smdistributed''' ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class _lowerCAmelCase ( __snake_case ): __lowerCAmelCase : str = field( default='''''' , metadata={'''help''': '''Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'''} , ) def _lowerCAmelCase ( self : List[str] ) -> int: """simple docstring""" super().__post_init__() warnings.warn( '''`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use ''' '''`TrainingArguments` instead.''' , a , ) @cached_property def _lowerCAmelCase ( self : Tuple ) -> "torch.device": """simple docstring""" logger.info('''PyTorch: setting up devices''' ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( '''torch.distributed process group is initialized, but local_rank == -1. ''' '''In order to use Torch DDP, launch your script with `python -m torch.distributed.launch''' ) if self.no_cuda: lowercase = torch.device('''cpu''' ) lowercase = 0 elif is_sagemaker_model_parallel_available(): lowercase = smp.local_rank() lowercase = torch.device('''cuda''' , a ) lowercase = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend='''smddp''' , timeout=self.ddp_timeout_delta ) lowercase = int(os.getenv('''SMDATAPARALLEL_LOCAL_RANK''' ) ) lowercase = torch.device('''cuda''' , self.local_rank ) lowercase = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 lowercase = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. lowercase = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend='''nccl''' , timeout=self.ddp_timeout_delta ) lowercase = torch.device('''cuda''' , self.local_rank ) lowercase = 1 if device.type == "cuda": torch.cuda.set_device(a ) return device @property def _lowerCAmelCase ( self : List[str] ) -> Dict: """simple docstring""" if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def _lowerCAmelCase ( self : str ) -> List[Any]: """simple docstring""" return not is_sagemaker_model_parallel_available() @property def _lowerCAmelCase ( self : str ) -> Dict: """simple docstring""" return False
396
1
import argparse import copy def lowercase__ ( A_: Tuple ) -> Dict: """simple docstring""" __UpperCAmelCase ={} with open(A_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __UpperCAmelCase =[] _list.append([line.split()[1], line.split()[2]] ) __UpperCAmelCase =_list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __UpperCAmelCase =[] _list.append([line.split()[0], line.split()[2]] ) __UpperCAmelCase =_list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def lowercase__ ( A_: str , A_: List[str] ) -> Optional[Any]: """simple docstring""" with open(A_ ) as f: __UpperCAmelCase =f.read(1 ) __UpperCAmelCase =start_node __UpperCAmelCase =[] __UpperCAmelCase =start_node __UpperCAmelCase =0 while visiting not in first_solution: __UpperCAmelCase =10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(A_ ) and k[0] not in first_solution: __UpperCAmelCase =k[1] __UpperCAmelCase =k[0] first_solution.append(A_ ) __UpperCAmelCase =distance_of_first_solution + int(A_ ) __UpperCAmelCase =best_node first_solution.append(A_ ) __UpperCAmelCase =0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __UpperCAmelCase =( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def lowercase__ ( A_: Any , A_: Optional[Any] ) -> str: """simple docstring""" __UpperCAmelCase =[] for n in solution[1:-1]: __UpperCAmelCase =solution.index(A_ ) for kn in solution[1:-1]: __UpperCAmelCase =solution.index(A_ ) if n == kn: continue __UpperCAmelCase =copy.deepcopy(A_ ) __UpperCAmelCase =kn __UpperCAmelCase =n __UpperCAmelCase =0 for k in _tmp[:-1]: __UpperCAmelCase =_tmp[_tmp.index(A_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __UpperCAmelCase =distance + int(i[1] ) _tmp.append(A_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __UpperCAmelCase =len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda A_ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def lowercase__ ( A_: List[Any] , A_: List[str] , A_: Union[str, Any] , A_: Optional[Any] , A_: Optional[Any] ) -> List[Any]: """simple docstring""" __UpperCAmelCase =1 __UpperCAmelCase =first_solution __UpperCAmelCase =[] __UpperCAmelCase =distance_of_first_solution __UpperCAmelCase =solution while count <= iters: __UpperCAmelCase =find_neighborhood(A_ , A_ ) __UpperCAmelCase =0 __UpperCAmelCase =neighborhood[index_of_best_solution] __UpperCAmelCase =len(A_ ) - 1 __UpperCAmelCase =False while not found: __UpperCAmelCase =0 while i < len(A_ ): if best_solution[i] != solution[i]: __UpperCAmelCase =best_solution[i] __UpperCAmelCase =solution[i] break __UpperCAmelCase =i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __UpperCAmelCase =True __UpperCAmelCase =best_solution[:-1] __UpperCAmelCase =neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __UpperCAmelCase =cost __UpperCAmelCase =solution else: __UpperCAmelCase =index_of_best_solution + 1 __UpperCAmelCase =neighborhood[index_of_best_solution] if len(A_ ) >= size: tabu_list.pop(0 ) __UpperCAmelCase =count + 1 return best_solution_ever, best_cost def lowercase__ ( A_: Union[str, Any]=None ) -> Tuple: """simple docstring""" __UpperCAmelCase =generate_neighbours(args.File ) __UpperCAmelCase , __UpperCAmelCase =generate_first_solution( args.File , A_ ) __UpperCAmelCase , __UpperCAmelCase =tabu_search( A_ , A_ , A_ , args.Iterations , args.Size , ) print(F'''Best solution: {best_sol}, with total distance: {best_cost}.''' ) if __name__ == "__main__": __A = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
68
'''simple docstring''' from __future__ import annotations def a ( UpperCamelCase_ : list[int] , UpperCamelCase_ : list[int] , UpperCamelCase_ : int ) -> tuple[float, list[float]]: snake_case__ =list(range(len(UpperCamelCase_ ) ) ) snake_case__ =[v / w for v, w in zip(UpperCamelCase_ , UpperCamelCase_ )] index.sort(key=lambda UpperCamelCase_ : ratio[i] , reverse=UpperCamelCase_ ) snake_case__ =0 snake_case__ =[0] * len(UpperCamelCase_ ) for i in index: if weight[i] <= capacity: snake_case__ =1 max_value += value[i] capacity -= weight[i] else: snake_case__ =capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
538
0
from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run UpperCamelCase_ = True except (ImportError, AttributeError): UpperCamelCase_ = object def _UpperCAmelCase ( *A , **A ): '''simple docstring''' pass UpperCamelCase_ = False UpperCamelCase_ = logging.get_logger('transformers-cli/serving') def _UpperCAmelCase ( A ): '''simple docstring''' UpperCAmelCase__ =pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(A , args.host , args.port , args.workers ) class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = 42 class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = 42 __UpperCamelCase = 42 class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = 42 class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = 42 class snake_case_ ( a ): '''simple docstring''' @staticmethod def __UpperCAmelCase ( A_ ) -> Optional[int]: UpperCAmelCase__ =parser.add_parser( "serve", help="CLI tool to run inference requests through REST and GraphQL endpoints." ) serve_parser.add_argument( "--task", type=A_, choices=get_supported_tasks(), help="The task to run the pipeline on", ) serve_parser.add_argument("--host", type=A_, default="localhost", help="Interface the server will listen on." ) serve_parser.add_argument("--port", type=A_, default=8888, help="Port the serving will listen to." ) serve_parser.add_argument("--workers", type=A_, default=1, help="Number of http workers" ) serve_parser.add_argument("--model", type=A_, help="Model's name or path to stored model." ) serve_parser.add_argument("--config", type=A_, help="Model's config name or path to stored model." ) serve_parser.add_argument("--tokenizer", type=A_, help="Tokenizer name to use." ) serve_parser.add_argument( "--device", type=A_, default=-1, help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)", ) serve_parser.set_defaults(func=A_ ) def __init__( self, A_, A_, A_, A_ ) -> str: UpperCAmelCase__ =pipeline UpperCAmelCase__ =host UpperCAmelCase__ =port UpperCAmelCase__ =workers if not _serve_dependencies_installed: raise RuntimeError( "Using serve command requires FastAPI and uvicorn. " "Please install transformers with [serving]: pip install \"transformers[serving]\"." "Or install FastAPI and uvicorn separately." ) else: logger.info(f"""Serving model over {host}:{port}""" ) UpperCAmelCase__ =FastAPI( routes=[ APIRoute( "/", self.model_info, response_model=A_, response_class=A_, methods=["GET"], ), APIRoute( "/tokenize", self.tokenize, response_model=A_, response_class=A_, methods=["POST"], ), APIRoute( "/detokenize", self.detokenize, response_model=A_, response_class=A_, methods=["POST"], ), APIRoute( "/forward", self.forward, response_model=A_, response_class=A_, methods=["POST"], ), ], timeout=600, ) def __UpperCAmelCase ( self ) -> Dict: run(self._app, host=self.host, port=self.port, workers=self.workers ) def __UpperCAmelCase ( self ) -> Any: return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def __UpperCAmelCase ( self, A_ = Body(A_, embed=A_ ), A_ = Body(A_, embed=A_ ) ) -> Optional[int]: try: UpperCAmelCase__ =self._pipeline.tokenizer.tokenize(A_ ) if return_ids: UpperCAmelCase__ =self._pipeline.tokenizer.convert_tokens_to_ids(A_ ) return ServeTokenizeResult(tokens=A_, tokens_ids=A_ ) else: return ServeTokenizeResult(tokens=A_ ) except Exception as e: raise HTTPException(status_code=500, detail={"model": "", "error": str(A_ )} ) def __UpperCAmelCase ( self, A_ = Body(A_, embed=A_ ), A_ = Body(A_, embed=A_ ), A_ = Body(A_, embed=A_ ), ) -> List[Any]: try: UpperCAmelCase__ =self._pipeline.tokenizer.decode(A_, A_, A_ ) return ServeDeTokenizeResult(model="", text=A_ ) except Exception as e: raise HTTPException(status_code=500, detail={"model": "", "error": str(A_ )} ) async def __UpperCAmelCase ( self, A_=Body(A_, embed=A_ ) ) -> int: # Check we don't have empty string if len(A_ ) == 0: return ServeForwardResult(output=[], attention=[] ) try: # Forward through the model UpperCAmelCase__ =self._pipeline(A_ ) return ServeForwardResult(output=A_ ) except Exception as e: raise HTTPException(500, {"error": str(A_ )} )
510
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { 'bigcode/gpt_bigcode-santacoder': 'https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json', } class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = 'gpt_bigcode' __UpperCamelCase = ['past_key_values'] __UpperCamelCase = { 'hidden_size': 'n_embd', 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self, A_=5_0257, A_=1024, A_=768, A_=12, A_=12, A_=None, A_="gelu_pytorch_tanh", A_=0.1, A_=0.1, A_=0.1, A_=1E-5, A_=0.02, A_=True, A_=True, A_=5_0256, A_=5_0256, A_=True, A_=True, A_=True, **A_, ) -> Dict: UpperCAmelCase__ =vocab_size UpperCAmelCase__ =n_positions UpperCAmelCase__ =n_embd UpperCAmelCase__ =n_layer UpperCAmelCase__ =n_head UpperCAmelCase__ =n_inner UpperCAmelCase__ =activation_function UpperCAmelCase__ =resid_pdrop UpperCAmelCase__ =embd_pdrop UpperCAmelCase__ =attn_pdrop UpperCAmelCase__ =layer_norm_epsilon UpperCAmelCase__ =initializer_range UpperCAmelCase__ =scale_attn_weights UpperCAmelCase__ =use_cache UpperCAmelCase__ =attention_softmax_in_fpaa UpperCAmelCase__ =scale_attention_softmax_in_fpaa UpperCAmelCase__ =multi_query UpperCAmelCase__ =bos_token_id UpperCAmelCase__ =eos_token_id super().__init__(bos_token_id=A_, eos_token_id=A_, **A_ )
510
1
'''simple docstring''' from ...processing_utils import ProcessorMixin class __A (lowerCamelCase__ ): snake_case :Any = "SpeechT5FeatureExtractor" snake_case :Optional[int] = "SpeechT5Tokenizer" def __init__( self , UpperCamelCase_ , UpperCamelCase_ ): super().__init__(lowercase_ , lowercase_ ) def __call__( self , *UpperCamelCase_ , **UpperCamelCase_ ): __UpperCAmelCase : int = kwargs.pop("audio" , lowercase_ ) __UpperCAmelCase : Optional[int] = kwargs.pop("text" , lowercase_ ) __UpperCAmelCase : List[Any] = kwargs.pop("text_target" , lowercase_ ) __UpperCAmelCase : List[str] = kwargs.pop("audio_target" , lowercase_ ) __UpperCAmelCase : List[Any] = kwargs.pop("sampling_rate" , lowercase_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __UpperCAmelCase : Dict = self.feature_extractor(lowercase_ , *lowercase_ , sampling_rate=lowercase_ , **lowercase_ ) elif text is not None: __UpperCAmelCase : Dict = self.tokenizer(lowercase_ , **lowercase_ ) else: __UpperCAmelCase : Optional[Any] = None if audio_target is not None: __UpperCAmelCase : Tuple = self.feature_extractor(audio_target=lowercase_ , *lowercase_ , sampling_rate=lowercase_ , **lowercase_ ) __UpperCAmelCase : List[Any] = targets["input_values"] elif text_target is not None: __UpperCAmelCase : List[str] = self.tokenizer(lowercase_ , **lowercase_ ) __UpperCAmelCase : List[Any] = targets["input_ids"] else: __UpperCAmelCase : List[Any] = None if inputs is None: return targets if targets is not None: __UpperCAmelCase : Any = labels __UpperCAmelCase : Any = targets.get("attention_mask" ) if decoder_attention_mask is not None: __UpperCAmelCase : List[str] = decoder_attention_mask return inputs def _snake_case ( self , *UpperCamelCase_ , **UpperCamelCase_ ): __UpperCAmelCase : Tuple = kwargs.pop("input_values" , lowercase_ ) __UpperCAmelCase : Optional[int] = kwargs.pop("input_ids" , lowercase_ ) __UpperCAmelCase : List[Any] = kwargs.pop("labels" , lowercase_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __UpperCAmelCase : Dict = self.feature_extractor.pad(lowercase_ , *lowercase_ , **lowercase_ ) elif input_ids is not None: __UpperCAmelCase : Union[str, Any] = self.tokenizer.pad(lowercase_ , **lowercase_ ) else: __UpperCAmelCase : str = None if labels is not None: if "input_ids" in labels or (isinstance(lowercase_ , lowercase_ ) and "input_ids" in labels[0]): __UpperCAmelCase : List[str] = self.tokenizer.pad(lowercase_ , **lowercase_ ) __UpperCAmelCase : List[str] = targets["input_ids"] else: __UpperCAmelCase : Tuple = self.feature_extractor.feature_size __UpperCAmelCase : str = self.feature_extractor.num_mel_bins __UpperCAmelCase : str = self.feature_extractor.pad(lowercase_ , *lowercase_ , **lowercase_ ) __UpperCAmelCase : Dict = feature_size_hack __UpperCAmelCase : Optional[Any] = targets["input_values"] else: __UpperCAmelCase : List[Any] = None if inputs is None: return targets if targets is not None: __UpperCAmelCase : Tuple = labels __UpperCAmelCase : Dict = targets.get("attention_mask" ) if decoder_attention_mask is not None: __UpperCAmelCase : Tuple = decoder_attention_mask return inputs def _snake_case ( self , *UpperCamelCase_ , **UpperCamelCase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def _snake_case ( self , *UpperCamelCase_ , **UpperCamelCase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ )
168
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer UpperCAmelCase__ : int = logging.get_logger(__name__) UpperCAmelCase__ : List[str] = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase__ : str = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } UpperCAmelCase__ : List[str] = { "distilbert-base-uncased": 5_12, "distilbert-base-uncased-distilled-squad": 5_12, "distilbert-base-cased": 5_12, "distilbert-base-cased-distilled-squad": 5_12, "distilbert-base-german-cased": 5_12, "distilbert-base-multilingual-cased": 5_12, } UpperCAmelCase__ : Dict = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class __lowercase ( lowerCamelCase__ ): __UpperCAmelCase = VOCAB_FILES_NAMES __UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase = PRETRAINED_INIT_CONFIGURATION __UpperCAmelCase = ['''input_ids''', '''attention_mask'''] __UpperCAmelCase = DistilBertTokenizer def __init__( self , lowercase_=None , lowercase_=None , lowercase_=True , lowercase_="[UNK]" , lowercase_="[SEP]" , lowercase_="[PAD]" , lowercase_="[CLS]" , lowercase_="[MASK]" , lowercase_=True , lowercase_=None , **lowercase_ , ) -> Any: super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , tokenize_chinese_chars=lowercase_ , strip_accents=lowercase_ , **lowercase_ , ) __snake_case = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('lowercase' , lowercase_) != do_lower_case or normalizer_state.get('strip_accents' , lowercase_) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowercase_) != tokenize_chinese_chars ): __snake_case = getattr(lowercase_ , normalizer_state.pop('type')) __snake_case = do_lower_case __snake_case = strip_accents __snake_case = tokenize_chinese_chars __snake_case = normalizer_class(**lowercase_) __snake_case = do_lower_case def _a ( self , lowercase_ , lowercase_=None) -> Union[str, Any]: __snake_case = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _a ( self , lowercase_ , lowercase_ = None) -> List[int]: __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _a ( self , lowercase_ , lowercase_ = None) -> Tuple[str]: __snake_case = self._tokenizer.model.save(lowercase_ , name=lowercase_) return tuple(lowercase_)
313
0
def __lowerCamelCase (UpperCAmelCase__ : list[int] ): if not numbers: return 0 if not isinstance(UpperCAmelCase__ , (list, tuple) ) or not all( isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) for number in numbers ): raise ValueError("numbers must be an iterable of integers" ) SCREAMING_SNAKE_CASE = SCREAMING_SNAKE_CASE = SCREAMING_SNAKE_CASE = numbers[0] for i in range(1 , len(UpperCAmelCase__ ) ): # update the maximum and minimum subarray products SCREAMING_SNAKE_CASE = numbers[i] if number < 0: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = min_till_now, max_till_now SCREAMING_SNAKE_CASE = max(UpperCAmelCase__ , max_till_now * number ) SCREAMING_SNAKE_CASE = min(UpperCAmelCase__ , min_till_now * number ) # update the maximum product found till now SCREAMING_SNAKE_CASE = max(UpperCAmelCase__ , UpperCAmelCase__ ) return max_prod
647
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) def __lowerCamelCase (UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] ): return [ int(1_0_0_0 * (box[0] / width) ), int(1_0_0_0 * (box[1] / height) ), int(1_0_0_0 * (box[2] / width) ), int(1_0_0_0 * (box[3] / height) ), ] def __lowerCamelCase (UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Optional[str] , UpperCAmelCase__ : Optional[str] = None ): SCREAMING_SNAKE_CASE = tesseract_config if tesseract_config is not None else "" # apply OCR SCREAMING_SNAKE_CASE = to_pil_image(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = pil_image.size SCREAMING_SNAKE_CASE = pytesseract.image_to_data(UpperCAmelCase__ , lang=UpperCAmelCase__ , output_type="dict" , config=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates SCREAMING_SNAKE_CASE = [idx for idx, word in enumerate(UpperCAmelCase__ ) if not word.strip()] SCREAMING_SNAKE_CASE = [word for idx, word in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE = [coord for idx, coord in enumerate(UpperCAmelCase__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format SCREAMING_SNAKE_CASE = [] for x, y, w, h in zip(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ): SCREAMING_SNAKE_CASE = [x, y, x + w, y + h] actual_boxes.append(UpperCAmelCase__ ) # finally, normalize the bounding boxes SCREAMING_SNAKE_CASE = [] for box in actual_boxes: normalized_boxes.append(normalize_box(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class lowercase ( a ): lowercase__ : Optional[int] = ["""pixel_values"""] def __init__( self : int , _UpperCamelCase : bool = True , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : PILImageResampling = PILImageResampling.BILINEAR , _UpperCamelCase : bool = True , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[str] = "" , **_UpperCamelCase : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = size if size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = resample SCREAMING_SNAKE_CASE = apply_ocr SCREAMING_SNAKE_CASE = ocr_lang SCREAMING_SNAKE_CASE = tesseract_config def __snake_case( self : List[Any] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Dict[str, int] , _UpperCamelCase : PILImageResampling = PILImageResampling.BILINEAR , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Any , ) -> np.ndarray: '''simple docstring''' SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) SCREAMING_SNAKE_CASE = (size["height"], size["width"]) return resize(_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : Tuple , _UpperCamelCase : ImageInput , _UpperCamelCase : bool = None , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : PILImageResampling = None , _UpperCamelCase : bool = None , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[str] = None , _UpperCamelCase : Optional[Union[str, TensorType]] = None , _UpperCamelCase : ChannelDimension = ChannelDimension.FIRST , **_UpperCamelCase : str , ) -> PIL.Image.Image: '''simple docstring''' SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE = size if size is not None else self.size SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase ) SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE = apply_ocr if apply_ocr is not None else self.apply_ocr SCREAMING_SNAKE_CASE = ocr_lang if ocr_lang is not None else self.ocr_lang SCREAMING_SNAKE_CASE = tesseract_config if tesseract_config is not None else self.tesseract_config SCREAMING_SNAKE_CASE = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(_UpperCamelCase ) for image in images] if apply_ocr: requires_backends(self , "pytesseract" ) SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for image in images: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = apply_tesseract(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) words_batch.append(_UpperCamelCase ) boxes_batch.append(_UpperCamelCase ) if do_resize: SCREAMING_SNAKE_CASE = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) SCREAMING_SNAKE_CASE = [flip_channel_order(_UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = BatchFeature(data={"pixel_values": images} , tensor_type=_UpperCamelCase ) if apply_ocr: SCREAMING_SNAKE_CASE = words_batch SCREAMING_SNAKE_CASE = boxes_batch return data
647
1
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class a ( UpperCAmelCase__ ): def __get__( self : Optional[int] , lowerCAmelCase : List[str] , lowerCAmelCase : str=None ) -> Optional[Any]: '''simple docstring''' if obj is None: return self if self.fget is None: raise AttributeError("""unreadable attribute""" ) SCREAMING_SNAKE_CASE_: Tuple ="""__cached_""" + self.fget.__name__ SCREAMING_SNAKE_CASE_: Union[str, Any] =getattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if cached is None: SCREAMING_SNAKE_CASE_: Tuple =self.fget(lowerCAmelCase ) setattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return cached def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Optional[int] =val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def __magic_name__ ( lowercase ): if is_torch_fx_proxy(lowercase ): return True if is_torch_available(): import torch if isinstance(lowercase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowercase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowercase , (jnp.ndarray, Tracer) ): return True return isinstance(lowercase , np.ndarray ) def __magic_name__ ( lowercase ): return isinstance(lowercase , np.ndarray ) def __magic_name__ ( lowercase ): return _is_numpy(lowercase ) def __magic_name__ ( lowercase ): import torch return isinstance(lowercase , torch.Tensor ) def __magic_name__ ( lowercase ): return False if not is_torch_available() else _is_torch(lowercase ) def __magic_name__ ( lowercase ): import torch return isinstance(lowercase , torch.device ) def __magic_name__ ( lowercase ): return False if not is_torch_available() else _is_torch_device(lowercase ) def __magic_name__ ( lowercase ): import torch if isinstance(lowercase , lowercase ): if hasattr(lowercase , lowercase ): SCREAMING_SNAKE_CASE_: str =getattr(lowercase , lowercase ) else: return False return isinstance(lowercase , torch.dtype ) def __magic_name__ ( lowercase ): return False if not is_torch_available() else _is_torch_dtype(lowercase ) def __magic_name__ ( lowercase ): import tensorflow as tf return isinstance(lowercase , tf.Tensor ) def __magic_name__ ( lowercase ): return False if not is_tf_available() else _is_tensorflow(lowercase ) def __magic_name__ ( lowercase ): import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowercase , """is_symbolic_tensor""" ): return tf.is_symbolic_tensor(lowercase ) return type(lowercase ) == tf.Tensor def __magic_name__ ( lowercase ): return False if not is_tf_available() else _is_tf_symbolic_tensor(lowercase ) def __magic_name__ ( lowercase ): import jax.numpy as jnp # noqa: F811 return isinstance(lowercase , jnp.ndarray ) def __magic_name__ ( lowercase ): return False if not is_flax_available() else _is_jax(lowercase ) def __magic_name__ ( lowercase ): if isinstance(lowercase , (dict, UserDict) ): return {k: to_py_obj(lowercase ) for k, v in obj.items()} elif isinstance(lowercase , (list, tuple) ): return [to_py_obj(lowercase ) for o in obj] elif is_tf_tensor(lowercase ): return obj.numpy().tolist() elif is_torch_tensor(lowercase ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowercase ): return np.asarray(lowercase ).tolist() elif isinstance(lowercase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def __magic_name__ ( lowercase ): if isinstance(lowercase , (dict, UserDict) ): return {k: to_numpy(lowercase ) for k, v in obj.items()} elif isinstance(lowercase , (list, tuple) ): return np.array(lowercase ) elif is_tf_tensor(lowercase ): return obj.numpy() elif is_torch_tensor(lowercase ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowercase ): return np.asarray(lowercase ) else: return obj class a ( UpperCAmelCase__ ): def lowerCamelCase__ ( self : Optional[int] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =fields(self ) # Safety and consistency checks if not len(lowerCAmelCase ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) SCREAMING_SNAKE_CASE_: Any =getattr(self , class_fields[0].name ) SCREAMING_SNAKE_CASE_: List[str] =all(getattr(self , field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCAmelCase ): if isinstance(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE_: str =first_field.items() SCREAMING_SNAKE_CASE_: Any =True else: try: SCREAMING_SNAKE_CASE_: Union[str, Any] =iter(lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Any =True except TypeError: SCREAMING_SNAKE_CASE_: Optional[int] =False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCAmelCase ): if ( not isinstance(lowerCAmelCase , (list, tuple) ) or not len(lowerCAmelCase ) == 2 or not isinstance(element[0] , lowerCAmelCase ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute SCREAMING_SNAKE_CASE_: int =first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self , element[0] , element[1] ) if element[1] is not None: SCREAMING_SNAKE_CASE_: str =element[1] elif first_field is not None: SCREAMING_SNAKE_CASE_: Tuple =first_field else: for field in class_fields: SCREAMING_SNAKE_CASE_: List[Any] =getattr(self , field.name ) if v is not None: SCREAMING_SNAKE_CASE_: Optional[Any] =v def __delitem__( self : Dict , *lowerCAmelCase : List[Any] , **lowerCAmelCase : Optional[Any] ) -> Optional[Any]: '''simple docstring''' raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def lowerCamelCase__ ( self : Dict , *lowerCAmelCase : Union[str, Any] , **lowerCAmelCase : Optional[int] ) -> str: '''simple docstring''' raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def lowerCamelCase__ ( self : int , *lowerCAmelCase : str , **lowerCAmelCase : Tuple ) -> Optional[int]: '''simple docstring''' raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def lowerCamelCase__ ( self : int , *lowerCAmelCase : str , **lowerCAmelCase : Union[str, Any] ) -> List[str]: '''simple docstring''' raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Union[str, Any] , lowerCAmelCase : int ) -> Tuple: '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] =dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : Tuple , lowerCAmelCase : Optional[Any] , lowerCAmelCase : List[str] ) -> Optional[int]: '''simple docstring''' if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCAmelCase , lowerCAmelCase ) super().__setattr__(lowerCAmelCase , lowerCAmelCase ) def __setitem__( self : Optional[int] , lowerCAmelCase : int , lowerCAmelCase : Union[str, Any] ) -> Optional[int]: '''simple docstring''' super().__setitem__(lowerCAmelCase , lowerCAmelCase ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCAmelCase , lowerCAmelCase ) def lowerCamelCase__ ( self : str ) -> Tuple[Any]: '''simple docstring''' return tuple(self[k] for k in self.keys() ) class a ( UpperCAmelCase__ , UpperCAmelCase__ ): @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , lowerCAmelCase : str ) -> Optional[int]: '''simple docstring''' raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class a ( UpperCAmelCase__ ): UpperCamelCase : Union[str, Any] = 'longest' UpperCamelCase : Any = 'max_length' UpperCamelCase : Optional[Any] = 'do_not_pad' class a ( UpperCAmelCase__ ): UpperCamelCase : Any = 'pt' UpperCamelCase : Optional[int] = 'tf' UpperCamelCase : Tuple = 'np' UpperCamelCase : str = 'jax' class a : def __init__( self : Dict , lowerCAmelCase : List[ContextManager] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE_: Any =context_managers SCREAMING_SNAKE_CASE_: List[Any] =ExitStack() def __enter__( self : Optional[int] ) -> Tuple: '''simple docstring''' for context_manager in self.context_managers: self.stack.enter_context(lowerCAmelCase ) def __exit__( self : List[Any] , *lowerCAmelCase : int , **lowerCAmelCase : Optional[Any] ) -> Dict: '''simple docstring''' self.stack.__exit__(*lowerCAmelCase , **lowerCAmelCase ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Dict =infer_framework(lowercase ) if framework == "tf": SCREAMING_SNAKE_CASE_: Tuple =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": SCREAMING_SNAKE_CASE_: Any =inspect.signature(model_class.forward ) # PyTorch models else: SCREAMING_SNAKE_CASE_: int =inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[str] =model_class.__name__ SCREAMING_SNAKE_CASE_: Tuple =infer_framework(lowercase ) if framework == "tf": SCREAMING_SNAKE_CASE_: List[str] =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": SCREAMING_SNAKE_CASE_: Tuple =inspect.signature(model_class.forward ) # PyTorch models else: SCREAMING_SNAKE_CASE_: List[Any] =inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def __magic_name__ ( lowercase , lowercase = "" , lowercase = "." ): def _flatten_dict(lowercase , lowercase="" , lowercase="." ): for k, v in d.items(): SCREAMING_SNAKE_CASE_: Optional[Any] =str(lowercase ) + delimiter + str(lowercase ) if parent_key else k if v and isinstance(lowercase , lowercase ): yield from flatten_dict(lowercase , lowercase , delimiter=lowercase ).items() else: yield key, v return dict(_flatten_dict(lowercase , lowercase , lowercase ) ) @contextmanager def __magic_name__ ( lowercase , lowercase = False ): if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def __magic_name__ ( lowercase , lowercase=None ): if is_numpy_array(lowercase ): return np.transpose(lowercase , axes=lowercase ) elif is_torch_tensor(lowercase ): return array.T if axes is None else array.permute(*lowercase ) elif is_tf_tensor(lowercase ): import tensorflow as tf return tf.transpose(lowercase , perm=lowercase ) elif is_jax_tensor(lowercase ): return jnp.transpose(lowercase , axes=lowercase ) else: raise ValueError(f'''Type not supported for transpose: {type(lowercase )}.''' ) def __magic_name__ ( lowercase , lowercase ): if is_numpy_array(lowercase ): return np.reshape(lowercase , lowercase ) elif is_torch_tensor(lowercase ): return array.reshape(*lowercase ) elif is_tf_tensor(lowercase ): import tensorflow as tf return tf.reshape(lowercase , lowercase ) elif is_jax_tensor(lowercase ): return jnp.reshape(lowercase , lowercase ) else: raise ValueError(f'''Type not supported for reshape: {type(lowercase )}.''' ) def __magic_name__ ( lowercase , lowercase=None ): if is_numpy_array(lowercase ): return np.squeeze(lowercase , axis=lowercase ) elif is_torch_tensor(lowercase ): return array.squeeze() if axis is None else array.squeeze(dim=lowercase ) elif is_tf_tensor(lowercase ): import tensorflow as tf return tf.squeeze(lowercase , axis=lowercase ) elif is_jax_tensor(lowercase ): return jnp.squeeze(lowercase , axis=lowercase ) else: raise ValueError(f'''Type not supported for squeeze: {type(lowercase )}.''' ) def __magic_name__ ( lowercase , lowercase ): if is_numpy_array(lowercase ): return np.expand_dims(lowercase , lowercase ) elif is_torch_tensor(lowercase ): return array.unsqueeze(dim=lowercase ) elif is_tf_tensor(lowercase ): import tensorflow as tf return tf.expand_dims(lowercase , axis=lowercase ) elif is_jax_tensor(lowercase ): return jnp.expand_dims(lowercase , axis=lowercase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(lowercase )}.''' ) def __magic_name__ ( lowercase ): if is_numpy_array(lowercase ): return np.size(lowercase ) elif is_torch_tensor(lowercase ): return array.numel() elif is_tf_tensor(lowercase ): import tensorflow as tf return tf.size(lowercase ) elif is_jax_tensor(lowercase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(lowercase )}.''' ) def __magic_name__ ( lowercase , lowercase ): for key, value in auto_map.items(): if isinstance(lowercase , (tuple, list) ): SCREAMING_SNAKE_CASE_: Dict =[f'''{repo_id}--{v}''' if (v is not None and """--""" not in v) else v for v in value] elif value is not None and "--" not in value: SCREAMING_SNAKE_CASE_: List[Any] =f'''{repo_id}--{value}''' return auto_map def __magic_name__ ( lowercase ): for base_class in inspect.getmro(lowercase ): SCREAMING_SNAKE_CASE_: str =base_class.__module__ SCREAMING_SNAKE_CASE_: Any =base_class.__name__ if module.startswith("""tensorflow""" ) or module.startswith("""keras""" ) or name == "TFPreTrainedModel": return "tf" elif module.startswith("""torch""" ) or name == "PreTrainedModel": return "pt" elif module.startswith("""flax""" ) or module.startswith("""jax""" ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
409
"""simple docstring""" from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch("""socket.socket""" ) @patch("""builtins.open""" ) def __magic_name__ ( lowercase , lowercase ): # ===== initialization ===== SCREAMING_SNAKE_CASE_: int =Mock() SCREAMING_SNAKE_CASE_: int =conn, Mock() SCREAMING_SNAKE_CASE_: Tuple =iter([1, None] ) SCREAMING_SNAKE_CASE_: Optional[Any] =lambda lowercase : next(lowercase ) # ===== invoke ===== send_file(filename="""mytext.txt""" , testing=lowercase ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
409
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:List[Any] = { """google/mobilenet_v2_1.4_224""": """https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json""", """google/mobilenet_v2_1.0_224""": """https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json""", """google/mobilenet_v2_0.75_160""": """https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json""", """google/mobilenet_v2_0.35_96""": """https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json""", # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class snake_case__ ( snake_case_ ): _snake_case : int = """mobilenet_v2""" def __init__( self , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=1.0 , lowerCamelCase=8 , lowerCamelCase=8 , lowerCamelCase=6 , lowerCamelCase=32 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="relu6" , lowerCamelCase=True , lowerCamelCase=0.8 , lowerCamelCase=0.02 , lowerCamelCase=0.001 , lowerCamelCase=255 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) __a = num_channels __a = image_size __a = depth_multiplier __a = depth_divisible_by __a = min_depth __a = expand_ratio __a = output_stride __a = first_layer_is_expansion __a = finegrained_output __a = hidden_act __a = tf_padding __a = classifier_dropout_prob __a = initializer_range __a = layer_norm_eps __a = semantic_loss_ignore_index class snake_case__ ( snake_case_ ): _snake_case : Union[str, Any] = version.parse("""1.11""" ) @property def a__ ( self ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def a__ ( self ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def a__ ( self ): return 1E-4
67
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __a = image_std if image_std is not None else OPENAI_CLIP_STD __a = do_convert_rgb def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __a = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , param_name="size" , default_to_square=lowerCamelCase ) __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" , default_to_square=lowerCamelCase ) __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __a = [convert_to_rgb(lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
1
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase : str = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class A_ ( _a , unittest.TestCase ): lowerCAmelCase__ = XLMRobertaTokenizer lowerCAmelCase__ = XLMRobertaTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def _lowercase ( self: Optional[int] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase : Optional[int] = XLMRobertaTokenizer(__lowerCAmelCase ,keep_accents=__lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self: Any ): '''simple docstring''' _lowerCamelCase : str = "<pad>" _lowerCamelCase : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCAmelCase ) ,__lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCAmelCase ) ,__lowerCAmelCase ) def _lowercase ( self: Union[str, Any] ): '''simple docstring''' _lowerCamelCase : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"<s>" ) self.assertEqual(vocab_keys[1] ,"<pad>" ) self.assertEqual(vocab_keys[-1] ,"<mask>" ) self.assertEqual(len(__lowerCAmelCase ) ,1_002 ) def _lowercase ( self: Union[str, Any] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,1_002 ) def _lowercase ( self: List[Any] ): '''simple docstring''' _lowerCamelCase : str = XLMRobertaTokenizer(__lowerCAmelCase ,keep_accents=__lowerCAmelCase ) _lowerCamelCase : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCAmelCase ,["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) _lowerCamelCase : Any = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCAmelCase ,[ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] ,) _lowerCamelCase : Optional[int] = tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) self.assertListEqual( __lowerCAmelCase ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] ,) _lowerCamelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowerCAmelCase ) self.assertListEqual( __lowerCAmelCase ,[ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] ,) def _lowercase ( self: Dict ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase : List[Any] = (self.rust_tokenizer_class, "hf-internal-testing/tiny-xlm-roberta", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase : Optional[int] = self.rust_tokenizer_class.from_pretrained(__lowerCAmelCase ,**__lowerCAmelCase ) _lowerCamelCase : str = self.tokenizer_class.from_pretrained(__lowerCAmelCase ,**__lowerCAmelCase ) _lowerCamelCase : List[Any] = tempfile.mkdtemp() _lowerCamelCase : Tuple = tokenizer_r.save_pretrained(__lowerCAmelCase ) _lowerCamelCase : Optional[Any] = tokenizer_p.save_pretrained(__lowerCAmelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) _lowerCamelCase : List[Any] = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(__lowerCAmelCase ,__lowerCAmelCase ) # Checks everything loads correctly in the same way _lowerCamelCase : Optional[int] = tokenizer_r.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Dict = tokenizer_p.from_pretrained(__lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase ,__lowerCAmelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowerCAmelCase ) # Save tokenizer rust, legacy_format=True _lowerCamelCase : List[str] = tempfile.mkdtemp() _lowerCamelCase : Optional[int] = tokenizer_r.save_pretrained(__lowerCAmelCase ,legacy_format=__lowerCAmelCase ) _lowerCamelCase : Any = tokenizer_p.save_pretrained(__lowerCAmelCase ) # Checks it save with the same files self.assertSequenceEqual(__lowerCAmelCase ,__lowerCAmelCase ) # Checks everything loads correctly in the same way _lowerCamelCase : Optional[Any] = tokenizer_r.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Tuple = tokenizer_p.from_pretrained(__lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase ,__lowerCAmelCase ) ) shutil.rmtree(__lowerCAmelCase ) # Save tokenizer rust, legacy_format=False _lowerCamelCase : Optional[Any] = tempfile.mkdtemp() _lowerCamelCase : List[str] = tokenizer_r.save_pretrained(__lowerCAmelCase ,legacy_format=__lowerCAmelCase ) _lowerCamelCase : List[Any] = tokenizer_p.save_pretrained(__lowerCAmelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase : Any = tokenizer_r.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Dict = tokenizer_p.from_pretrained(__lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase ,__lowerCAmelCase ) ) shutil.rmtree(__lowerCAmelCase ) @cached_property def _lowercase ( self: List[str] ): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained("xlm-roberta-base" ) def _lowercase ( self: int ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__lowerCAmelCase ,f.name ) _lowerCamelCase : List[Any] = XLMRobertaTokenizer(f.name ,keep_accents=__lowerCAmelCase ) _lowerCamelCase : List[Any] = pickle.dumps(__lowerCAmelCase ) pickle.loads(__lowerCAmelCase ) def _lowercase ( self: Optional[Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _lowerCamelCase : int = self.get_tokenizer() _lowerCamelCase : Union[str, Any] = self.get_rust_tokenizer() _lowerCamelCase : Optional[Any] = "I was born in 92000, and this is falsé." _lowerCamelCase : Union[str, Any] = tokenizer.tokenize(__lowerCAmelCase ) _lowerCamelCase : Optional[int] = rust_tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase ,__lowerCAmelCase ) _lowerCamelCase : Optional[Any] = tokenizer.encode(__lowerCAmelCase ,add_special_tokens=__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = rust_tokenizer.encode(__lowerCAmelCase ,add_special_tokens=__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase ,__lowerCAmelCase ) _lowerCamelCase : Tuple = self.get_rust_tokenizer() _lowerCamelCase : List[str] = tokenizer.encode(__lowerCAmelCase ) _lowerCamelCase : List[str] = rust_tokenizer.encode(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase ,__lowerCAmelCase ) @slow def _lowercase ( self: Dict ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = "Hello World!" _lowerCamelCase : Optional[Any] = [0, 35_378, 6_661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCAmelCase ,self.big_tokenizer.encode(__lowerCAmelCase ) ) @slow def _lowercase ( self: Tuple ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _lowerCamelCase : Any = [ 0, 3_293, 83, 10, 4_552, 4_989, 7_986, 678, 10, 5_915, 111, 179_459, 124_850, 4, 6_044, 237, 12, 6, 5, 6, 4, 6_780, 705, 15, 1_388, 44, 378, 10_114, 711, 152, 20, 6, 5, 22_376, 642, 1_221, 15_190, 34_153, 450, 5_608, 959, 1_119, 57_702, 136, 186, 47, 1_098, 29_367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_044, 237, 6_284, 50_901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCAmelCase ,self.big_tokenizer.encode(__lowerCAmelCase ) ) @slow def _lowercase ( self: int ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = {"input_ids": [[0, 11_062, 82_772, 7, 15, 82_772, 538, 51_529, 237, 17_198, 1_290, 206, 9, 215_175, 1_314, 136, 17_198, 1_290, 206, 9, 56_359, 42, 122_009, 9, 16_466, 16, 87_344, 4_537, 9, 4_717, 78_381, 6, 159_958, 7, 15, 24_480, 618, 4, 527, 22_693, 5_428, 4, 2_777, 24_480, 9_874, 4, 43_523, 594, 4, 803, 18_392, 33_189, 18, 4, 43_523, 24_447, 12_399, 100, 24_955, 83_658, 9_626, 144_057, 15, 839, 22_335, 16, 136, 24_955, 83_658, 83_479, 15, 39_102, 724, 16, 678, 645, 2_789, 1_328, 4_589, 42, 122_009, 115_774, 23, 805, 1_328, 46_876, 7, 136, 53_894, 1_940, 42_227, 41_159, 17_721, 823, 425, 4, 27_512, 98_722, 206, 136, 5_531, 4_970, 919, 17_336, 5, 2], [0, 20_080, 618, 83, 82_775, 47, 479, 9, 1_517, 73, 53_894, 333, 80_581, 110_117, 18_811, 5_256, 1_295, 51, 152_526, 297, 7_986, 390, 124_416, 538, 35_431, 214, 98, 15_044, 25_737, 136, 7_108, 43_701, 23, 756, 135_355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63_773, 119_455, 6, 147_797, 88_203, 7, 645, 70, 21, 3_285, 10_269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCAmelCase ,model_name="xlm-roberta-base" ,revision="d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3" ,)
46
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class lowercase__ : """simple docstring""" def __init__( self : Dict , __a : List[str] , __a : List[str]=1_3 , __a : List[str]=7 , __a : Dict=True , __a : str=True , __a : str=9_9 , __a : Dict=3_2 , __a : Optional[int]=5 , __a : List[Any]=4 , __a : Dict=3_7 , __a : List[Any]="gelu" , __a : str=0.1 , __a : Dict=0.1 , __a : Optional[Any]=5_0 , __a : Dict=0.02 , __a : List[Any]=True , __a : str=None , ): snake_case__ : int = parent snake_case__ : Any = batch_size snake_case__ : Any = seq_length snake_case__ : Dict = is_training snake_case__ : str = use_input_mask snake_case__ : Optional[Any] = vocab_size snake_case__ : List[Any] = hidden_size snake_case__ : Any = num_hidden_layers snake_case__ : Any = num_attention_heads snake_case__ : Optional[Any] = intermediate_size snake_case__ : str = hidden_act snake_case__ : Any = hidden_dropout_prob snake_case__ : List[str] = attention_probs_dropout_prob snake_case__ : Any = max_position_embeddings snake_case__ : Tuple = initializer_range snake_case__ : str = use_labels snake_case__ : List[str] = scope def lowercase ( self : Optional[Any] ): snake_case__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ : Any = None if self.use_input_mask: snake_case__ : int = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: snake_case__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ : Optional[int] = self.get_config() return config, input_ids, input_mask, token_labels def lowercase ( self : Optional[Any] ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=__a , initializer_range=self.initializer_range , ) def lowercase ( self : List[str] ): ( ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ) : Any = self.prepare_config_and_inputs() snake_case__ : Union[str, Any] = True snake_case__ : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def lowercase ( self : int , __a : List[Any] , __a : Optional[Any] , __a : Tuple , __a : Optional[int] , **__a : Union[str, Any] , ): snake_case__ : Optional[Any] = BertGenerationEncoder(config=__a ) model.to(__a ) model.eval() snake_case__ : Optional[int] = model(__a , attention_mask=__a ) snake_case__ : str = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase ( self : str , __a : Union[str, Any] , __a : Any , __a : int , __a : int , __a : List[Any] , __a : List[str] , **__a : List[Any] , ): snake_case__ : Union[str, Any] = True snake_case__ : int = BertGenerationEncoder(config=__a ) model.to(__a ) model.eval() snake_case__ : List[str] = model( __a , attention_mask=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , ) snake_case__ : List[Any] = model( __a , attention_mask=__a , encoder_hidden_states=__a , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase ( self : int , __a : Tuple , __a : Union[str, Any] , __a : Optional[Any] , __a : str , __a : Union[str, Any] , __a : str , **__a : List[Any] , ): snake_case__ : int = True snake_case__ : Optional[int] = True snake_case__ : List[str] = BertGenerationDecoder(config=__a ).to(__a ).eval() # first forward pass snake_case__ : Union[str, Any] = model( __a , attention_mask=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , use_cache=__a , ) snake_case__ : List[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case__ : Tuple = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case__ : int = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case__ : Tuple = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case__ : Any = model( __a , attention_mask=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , output_hidden_states=__a , )["""hidden_states"""][0] snake_case__ : Any = model( __a , attention_mask=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , past_key_values=__a , output_hidden_states=__a , )["""hidden_states"""][0] # select random slice snake_case__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case__ : Optional[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__a , __a , atol=1e-3 ) ) def lowercase ( self : Union[str, Any] , __a : List[Any] , __a : int , __a : Union[str, Any] , __a : int , *__a : str , ): snake_case__ : Union[str, Any] = BertGenerationDecoder(__a ) model.to(__a ) model.eval() snake_case__ : List[Any] = model(__a , attention_mask=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase ( self : Any ): snake_case__ , snake_case__ , snake_case__ , snake_case__ : str = self.prepare_config_and_inputs() snake_case__ : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowercase__ (__snake_case , __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Optional[Any] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () __UpperCamelCase : str = (BertGenerationDecoder,) if is_torch_available() else () __UpperCamelCase : str = ( {'feature-extraction': BertGenerationEncoder, 'text-generation': BertGenerationDecoder} if is_torch_available() else {} ) def lowercase ( self : str ): snake_case__ : Dict = BertGenerationEncoderTester(self ) snake_case__ : Tuple = ConfigTester(self , config_class=__a , hidden_size=3_7 ) def lowercase ( self : Optional[int] ): self.config_tester.run_common_tests() def lowercase ( self : Optional[int] ): snake_case__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def lowercase ( self : List[Any] ): snake_case__ , snake_case__ , snake_case__ , snake_case__ : Optional[int] = self.model_tester.prepare_config_and_inputs() snake_case__ : Optional[int] = """bert""" self.model_tester.create_and_check_model(__a , __a , __a , __a ) def lowercase ( self : str ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__a ) def lowercase ( self : Any ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__a ) def lowercase ( self : Tuple ): # This regression test was failing with PyTorch < 1.3 ( ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ) : int = self.model_tester.prepare_config_and_inputs_for_decoder() snake_case__ : int = None self.model_tester.create_and_check_model_as_decoder( __a , __a , __a , __a , __a , __a , ) def lowercase ( self : int ): snake_case__ : int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*__a ) @slow def lowercase ( self : List[str] ): snake_case__ : int = BertGenerationEncoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) self.assertIsNotNone(__a ) @require_torch class lowercase__ (unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : Optional[int] ): snake_case__ : Dict = BertGenerationEncoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) snake_case__ : Optional[Any] = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): snake_case__ : Union[str, Any] = model(__a )[0] snake_case__ : Optional[Any] = torch.Size([1, 8, 1_0_2_4] ) self.assertEqual(output.shape , __a ) snake_case__ : List[Any] = torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4 ) ) @require_torch class lowercase__ (unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : Any ): snake_case__ : List[str] = BertGenerationDecoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) snake_case__ : Tuple = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): snake_case__ : Any = model(__a )[0] snake_case__ : Optional[Any] = torch.Size([1, 8, 5_0_3_5_8] ) self.assertEqual(output.shape , __a ) snake_case__ : int = torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4 ) )
648
0
import argparse import copy def lowerCamelCase ( UpperCamelCase : List[Any] ) -> int: _lowerCamelCase = {} with open(UpperCamelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _lowerCamelCase = [] _list.append([line.split()[1], line.split()[2]] ) _lowerCamelCase = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _lowerCamelCase = [] _list.append([line.split()[0], line.split()[2]] ) _lowerCamelCase = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def lowerCamelCase ( UpperCamelCase : List[Any] , UpperCamelCase : List[Any] ) -> int: with open(UpperCamelCase ) as f: _lowerCamelCase = f.read(1 ) _lowerCamelCase = start_node _lowerCamelCase = [] _lowerCamelCase = start_node _lowerCamelCase = 0 while visiting not in first_solution: _lowerCamelCase = 1_00_00 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(UpperCamelCase ) and k[0] not in first_solution: _lowerCamelCase = k[1] _lowerCamelCase = k[0] first_solution.append(UpperCamelCase ) _lowerCamelCase = distance_of_first_solution + int(UpperCamelCase ) _lowerCamelCase = best_node first_solution.append(UpperCamelCase ) _lowerCamelCase = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _lowerCamelCase = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_00_00 ) return first_solution, distance_of_first_solution def lowerCamelCase ( UpperCamelCase : int , UpperCamelCase : Tuple ) -> List[Any]: _lowerCamelCase = [] for n in solution[1:-1]: _lowerCamelCase = solution.index(UpperCamelCase ) for kn in solution[1:-1]: _lowerCamelCase = solution.index(UpperCamelCase ) if n == kn: continue _lowerCamelCase = copy.deepcopy(UpperCamelCase ) _lowerCamelCase = kn _lowerCamelCase = n _lowerCamelCase = 0 for k in _tmp[:-1]: _lowerCamelCase = _tmp[_tmp.index(UpperCamelCase ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _lowerCamelCase = distance + int(i[1] ) _tmp.append(UpperCamelCase ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _lowerCamelCase = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda UpperCamelCase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def lowerCamelCase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : List[str] , UpperCamelCase : int , UpperCamelCase : Union[str, Any] , UpperCamelCase : Tuple ) -> Optional[int]: _lowerCamelCase = 1 _lowerCamelCase = first_solution _lowerCamelCase = [] _lowerCamelCase = distance_of_first_solution _lowerCamelCase = solution while count <= iters: _lowerCamelCase = find_neighborhood(UpperCamelCase , UpperCamelCase ) _lowerCamelCase = 0 _lowerCamelCase = neighborhood[index_of_best_solution] _lowerCamelCase = len(UpperCamelCase ) - 1 _lowerCamelCase = False while not found: _lowerCamelCase = 0 while i < len(UpperCamelCase ): if best_solution[i] != solution[i]: _lowerCamelCase = best_solution[i] _lowerCamelCase = solution[i] break _lowerCamelCase = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _lowerCamelCase = True _lowerCamelCase = best_solution[:-1] _lowerCamelCase = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _lowerCamelCase = cost _lowerCamelCase = solution else: _lowerCamelCase = index_of_best_solution + 1 _lowerCamelCase = neighborhood[index_of_best_solution] if len(UpperCamelCase ) >= size: tabu_list.pop(0 ) _lowerCamelCase = count + 1 return best_solution_ever, best_cost def lowerCamelCase ( UpperCamelCase : str=None ) -> int: _lowerCamelCase = generate_neighbours(args.File ) _lowerCamelCase , _lowerCamelCase = generate_first_solution( args.File , UpperCamelCase ) _lowerCamelCase , _lowerCamelCase = tabu_search( UpperCamelCase , UpperCamelCase , UpperCamelCase , args.Iterations , args.Size , ) print(F"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": A = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
234
from itertools import product def lowerCamelCase ( UpperCamelCase : int , UpperCamelCase : int ) -> list[int]: _lowerCamelCase = sides_number _lowerCamelCase = max_face_number * dice_number _lowerCamelCase = [0] * (max_total + 1) _lowerCamelCase = 1 _lowerCamelCase = range(UpperCamelCase , max_face_number + 1 ) for dice_numbers in product(UpperCamelCase , repeat=UpperCamelCase ): _lowerCamelCase = sum(UpperCamelCase ) totals_frequencies[total] += 1 return totals_frequencies def lowerCamelCase ( ) -> float: _lowerCamelCase = total_frequency_distribution( sides_number=4 , dice_number=9 ) _lowerCamelCase = total_frequency_distribution( sides_number=6 , dice_number=6 ) _lowerCamelCase = 0 _lowerCamelCase = 9 _lowerCamelCase = 4 * 9 _lowerCamelCase = 6 for peter_total in range(UpperCamelCase , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) _lowerCamelCase = (4**9) * (6**6) _lowerCamelCase = peter_wins_count / total_games_number _lowerCamelCase = round(UpperCamelCase , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(F'''{solution() = }''')
234
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: int = '''gpt_bigcode''' _lowerCamelCase: Tuple = ['''past_key_values'''] _lowerCamelCase: Dict = { '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : List[Any] ,A_ : str=5_0257 ,A_ : Optional[Any]=1024 ,A_ : Optional[int]=768 ,A_ : int=12 ,A_ : Dict=12 ,A_ : str=None ,A_ : Any="gelu_pytorch_tanh" ,A_ : int=0.1 ,A_ : Any=0.1 ,A_ : Tuple=0.1 ,A_ : str=1e-5 ,A_ : Optional[int]=0.02 ,A_ : Optional[Any]=True ,A_ : str=True ,A_ : Union[str, Any]=5_0256 ,A_ : Tuple=5_0256 ,A_ : Any=True ,A_ : Tuple=True ,A_ : int=True ,**A_ : Dict ,) -> Optional[Any]: A = vocab_size A = n_positions A = n_embd A = n_layer A = n_head A = n_inner A = activation_function A = resid_pdrop A = embd_pdrop A = attn_pdrop A = layer_norm_epsilon A = initializer_range A = scale_attn_weights A = use_cache A = attention_softmax_in_fpaa A = scale_attention_softmax_in_fpaa A = multi_query A = bos_token_id A = eos_token_id super().__init__(bos_token_id=A_ ,eos_token_id=A_ ,**A_ )
91
import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 lowerCamelCase = get_tests_dir('fixtures/dummy_feature_extractor_config.json') lowerCamelCase = get_tests_dir('fixtures/vocab.json') lowerCamelCase = get_tests_dir('fixtures') class A ( unittest.TestCase ): UpperCamelCase__ : Dict =['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] def lowerCamelCase ( self : str ) -> Optional[Any]: """simple docstring""" _lowerCamelCase : Dict =0 def lowerCamelCase ( self : str ) -> List[str]: """simple docstring""" _lowerCamelCase : List[Any] =AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : int ) -> Tuple: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : int =WavaVecaConfig() _lowerCamelCase : Dict =AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) # save in new folder model_config.save_pretrained(lowercase_ ) processor.save_pretrained(lowercase_ ) _lowerCamelCase : Any =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : str ) -> Union[str, Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(lowercase_ , os.path.join(lowercase_ , lowercase_ ) ) copyfile(lowercase_ , os.path.join(lowercase_ , 'vocab.json' ) ) _lowerCamelCase : Union[str, Any] =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : Any ) -> Optional[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : List[Any] =WavaVecaFeatureExtractor() _lowerCamelCase : List[str] =AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) _lowerCamelCase : str =WavaVecaProcessor(lowercase_ , lowercase_ ) # save in new folder processor.save_pretrained(lowercase_ ) # drop `processor_class` in tokenizer with open(os.path.join(lowercase_ , lowercase_ ) , 'r' ) as f: _lowerCamelCase : Optional[int] =json.load(lowercase_ ) config_dict.pop('processor_class' ) with open(os.path.join(lowercase_ , lowercase_ ) , 'w' ) as f: f.write(json.dumps(lowercase_ ) ) _lowerCamelCase : Optional[int] =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : Dict ) -> Optional[int]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Optional[Any] =WavaVecaFeatureExtractor() _lowerCamelCase : Tuple =AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) _lowerCamelCase : Dict =WavaVecaProcessor(lowercase_ , lowercase_ ) # save in new folder processor.save_pretrained(lowercase_ ) # drop `processor_class` in feature extractor with open(os.path.join(lowercase_ , lowercase_ ) , 'r' ) as f: _lowerCamelCase : Union[str, Any] =json.load(lowercase_ ) config_dict.pop('processor_class' ) with open(os.path.join(lowercase_ , lowercase_ ) , 'w' ) as f: f.write(json.dumps(lowercase_ ) ) _lowerCamelCase : Optional[int] =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Optional[Any] =WavaVecaConfig(processor_class='Wav2Vec2Processor' ) model_config.save_pretrained(lowercase_ ) # copy relevant files copyfile(lowercase_ , os.path.join(lowercase_ , 'vocab.json' ) ) # create emtpy sample processor with open(os.path.join(lowercase_ , lowercase_ ) , 'w' ) as f: f.write('{}' ) _lowerCamelCase : int =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def lowerCamelCase ( self : str ) -> Optional[int]: """simple docstring""" with self.assertRaises(lowercase_ ): _lowerCamelCase : int =AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowercase_ ): _lowerCamelCase : Union[str, Any] =AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowercase_ ) _lowerCamelCase : List[str] =AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowercase_ ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) _lowerCamelCase : int =processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) _lowerCamelCase : Optional[int] =processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version _lowerCamelCase : int =AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowercase_ , use_fast=lowercase_ ) _lowerCamelCase : Optional[int] =new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def lowerCamelCase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" try: AutoConfig.register('custom' , lowercase_ ) AutoFeatureExtractor.register(lowercase_ , lowercase_ ) AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) AutoProcessor.register(lowercase_ , lowercase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase_ ): AutoProcessor.register(lowercase_ , lowercase_ ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : str =CustomFeatureExtractor.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase : str =os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _lowerCamelCase : List[Any] =CustomTokenizer(lowercase_ ) _lowerCamelCase : Optional[int] =CustomProcessor(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(lowercase_ ) _lowerCamelCase : List[Any] =AutoProcessor.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" class A ( UpperCamelCase_ ): UpperCamelCase__ : Optional[Any] =False class A ( UpperCamelCase_ ): UpperCamelCase__ : int =False class A ( UpperCamelCase_ ): UpperCamelCase__ : Union[str, Any] ='AutoFeatureExtractor' UpperCamelCase__ : str ='AutoTokenizer' UpperCamelCase__ : List[Any] =False try: AutoConfig.register('custom' , lowercase_ ) AutoFeatureExtractor.register(lowercase_ , lowercase_ ) AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) AutoProcessor.register(lowercase_ , lowercase_ ) # If remote code is not set, the default is to use local classes. _lowerCamelCase : int =AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. _lowerCamelCase : int =AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowercase_ ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. _lowerCamelCase : str =AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowercase_ ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase ( self : Optional[int] ) -> str: """simple docstring""" _lowerCamelCase : List[Any] =AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(processor.__class__.__name__ , 'BertTokenizerFast' ) def lowerCamelCase ( self : Any ) -> Dict: """simple docstring""" _lowerCamelCase : Any =AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-convnext' ) self.assertEqual(processor.__class__.__name__ , 'ConvNextImageProcessor' ) @is_staging_test class A ( unittest.TestCase ): UpperCamelCase__ : List[Any] =['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def lowerCamelCase ( cls : int ) -> Optional[Any]: """simple docstring""" _lowerCamelCase : Union[str, Any] =TOKEN HfFolder.save_token(lowercase_ ) @classmethod def lowerCamelCase ( cls : Optional[int] ) -> Union[str, Any]: """simple docstring""" try: delete_repo(token=cls._token , repo_id='test-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-processor' ) except HTTPError: pass def lowerCamelCase ( self : str ) -> int: """simple docstring""" _lowerCamelCase : Tuple =WavaVecaProcessor.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(lowercase_ , 'test-processor' ) , push_to_hub=lowercase_ , use_auth_token=self._token ) _lowerCamelCase : Union[str, Any] =WavaVecaProcessor.from_pretrained(F'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(lowercase_ , getattr(new_processor.feature_extractor , lowercase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCamelCase ( self : str ) -> Tuple: """simple docstring""" _lowerCamelCase : int =WavaVecaProcessor.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(lowercase_ , 'test-processor-org' ) , push_to_hub=lowercase_ , use_auth_token=self._token , organization='valid_org' , ) _lowerCamelCase : str =WavaVecaProcessor.from_pretrained('valid_org/test-processor-org' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(lowercase_ , getattr(new_processor.feature_extractor , lowercase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCamelCase ( self : List[Any] ) -> str: """simple docstring""" CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() _lowerCamelCase : Optional[Any] =CustomFeatureExtractor.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase : Dict =os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _lowerCamelCase : Any =CustomTokenizer(lowercase_ ) _lowerCamelCase : List[Any] =CustomProcessor(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(F'''{USER}/test-dynamic-processor''' , token=self._token ) _lowerCamelCase : List[str] =Repository(lowercase_ , clone_from=F'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(lowercase_ ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { 'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor', 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(lowercase_ , 'tokenizer_config.json' ) ) as f: _lowerCamelCase : Union[str, Any] =json.load(lowercase_ ) self.assertDictEqual( tokenizer_config['auto_map'] , { 'AutoTokenizer': ['custom_tokenization.CustomTokenizer', None], 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(lowercase_ , 'custom_feature_extraction.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(lowercase_ , 'custom_tokenization.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(lowercase_ , 'custom_processing.py' ) ) ) repo.push_to_hub() _lowerCamelCase : Tuple =AutoProcessor.from_pretrained(F'''{USER}/test-dynamic-processor''' , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , 'CustomProcessor' )
464
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { """alibaba-damo/mgp-str-base""": """https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json""", } class lowerCAmelCase ( __UpperCAmelCase ): a : Optional[Any] = "mgp-str" def __init__( self , UpperCamelCase=[32, 128] , UpperCamelCase=4 , UpperCamelCase=3 , UpperCamelCase=27 , UpperCamelCase=38 , UpperCamelCase=50_257 , UpperCamelCase=30_522 , UpperCamelCase=768 , UpperCamelCase=12 , UpperCamelCase=12 , UpperCamelCase=4.0 , UpperCamelCase=True , UpperCamelCase=False , UpperCamelCase=1e-5 , UpperCamelCase=0.0 , UpperCamelCase=0.0 , UpperCamelCase=0.0 , UpperCamelCase=False , UpperCamelCase=0.02 , **UpperCamelCase , ): super().__init__(**UpperCamelCase ) _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = patch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = max_token_length _SCREAMING_SNAKE_CASE = num_character_labels _SCREAMING_SNAKE_CASE = num_bpe_labels _SCREAMING_SNAKE_CASE = num_wordpiece_labels _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = mlp_ratio _SCREAMING_SNAKE_CASE = distilled _SCREAMING_SNAKE_CASE = layer_norm_eps _SCREAMING_SNAKE_CASE = drop_rate _SCREAMING_SNAKE_CASE = qkv_bias _SCREAMING_SNAKE_CASE = attn_drop_rate _SCREAMING_SNAKE_CASE = drop_path_rate _SCREAMING_SNAKE_CASE = output_aa_attentions _SCREAMING_SNAKE_CASE = initializer_range
706
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class lowerCAmelCase ( __UpperCAmelCase ): a : Dict = DistilBertTokenizer a : Tuple = DistilBertTokenizerFast a : List[str] = True @slow def lowercase ( self ): _SCREAMING_SNAKE_CASE = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _SCREAMING_SNAKE_CASE = tokenizer.encode("sequence builders" , add_special_tokens=UpperCamelCase ) _SCREAMING_SNAKE_CASE = tokenizer.encode("multi-sequence build" , add_special_tokens=UpperCamelCase ) _SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCamelCase ) _SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCamelCase , UpperCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
493
0
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: int ): if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) SCREAMING_SNAKE_CASE__ = str(bin(UpperCamelCase__ ) )[2:] # remove the leading "0b" SCREAMING_SNAKE_CASE__ = str(bin(UpperCamelCase__ ) )[2:] SCREAMING_SNAKE_CASE__ = max(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) return "0b" + "".join( str(int("""1""" in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(UpperCamelCase__ ) , b_binary.zfill(UpperCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
6
import argparse import datetime def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = { """0""": """Sunday""", """1""": """Monday""", """2""": """Tuesday""", """3""": """Wednesday""", """4""": """Thursday""", """5""": """Friday""", """6""": """Saturday""", } SCREAMING_SNAKE_CASE__ = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(UpperCamelCase__ ) < 11: raise ValueError("""Must be 10 characters long""" ) # Get month SCREAMING_SNAKE_CASE__ = int(date_input[0] + date_input[1] ) # Validate if not 0 < m < 13: raise ValueError("""Month must be between 1 - 12""" ) SCREAMING_SNAKE_CASE__ = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError("""Date separator must be '-' or '/'""" ) # Get day SCREAMING_SNAKE_CASE__ = int(date_input[3] + date_input[4] ) # Validate if not 0 < d < 32: raise ValueError("""Date must be between 1 - 31""" ) # Get second separator SCREAMING_SNAKE_CASE__ = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError("""Date separator must be '-' or '/'""" ) # Get year SCREAMING_SNAKE_CASE__ = int(date_input[6] + date_input[7] + date_input[8] + date_input[9] ) # Arbitrary year range if not 45 < y < 8_500: raise ValueError( """Year out of range. There has to be some sort of limit...right?""" ) # Get datetime obj for validation SCREAMING_SNAKE_CASE__ = datetime.date(int(UpperCamelCase__ ) , int(UpperCamelCase__ ) , int(UpperCamelCase__ ) ) # Start math if m <= 2: SCREAMING_SNAKE_CASE__ = y - 1 SCREAMING_SNAKE_CASE__ = m + 12 # maths var SCREAMING_SNAKE_CASE__ = int(str(UpperCamelCase__ )[:2] ) SCREAMING_SNAKE_CASE__ = int(str(UpperCamelCase__ )[2:] ) SCREAMING_SNAKE_CASE__ = int(2.6 * m - 5.3_9 ) SCREAMING_SNAKE_CASE__ = int(c / 4 ) SCREAMING_SNAKE_CASE__ = int(k / 4 ) SCREAMING_SNAKE_CASE__ = int(d + k ) SCREAMING_SNAKE_CASE__ = int(t + u + v + x ) SCREAMING_SNAKE_CASE__ = int(z - (2 * c) ) SCREAMING_SNAKE_CASE__ = round(w % 7 ) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError("""The date was evaluated incorrectly. Contact developer.""" ) # Response SCREAMING_SNAKE_CASE__ = f'''Your date {date_input}, is a {days[str(UpperCamelCase__ )]}!''' return response if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase = argparse.ArgumentParser( description=( 'Find out what day of the week nearly any date is or was. Enter ' 'date as a string in the mm-dd-yyyy or mm/dd/yyyy format' ) ) parser.add_argument( 'date_input', type=str, help='Date as a string (mm-dd-yyyy or mm/dd/yyyy)' ) _lowerCamelCase = parser.parse_args() zeller(args.date_input)
6
1
_lowercase = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" _lowercase = [{"type": "code", "content": INSTALL_CONTENT}] _lowercase = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
708
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( )-> str: A__ = 1_0 A__ = datasets.Features( { '''tokens''': datasets.Sequence(datasets.Value('''string''' ) ), '''labels''': datasets.Sequence(datasets.ClassLabel(names=['''negative''', '''positive'''] ) ), '''answers''': datasets.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), '''id''': datasets.Value('''int64''' ), } ) A__ = datasets.Dataset.from_dict( { '''tokens''': [['''foo'''] * 5] * n, '''labels''': [[1] * 5] * n, '''answers''': [{'''answer_start''': [9_7], '''text''': ['''1976''']}] * 1_0, '''id''': list(range(UpperCamelCase_ ) ), } , features=UpperCamelCase_ , ) return dataset @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : int )-> Optional[Any]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''file.arrow''' ) dataset.map(cache_file_name=UpperCamelCase_ ) return filename # FILE_CONTENT + files _lowercase = "\\n Text data.\n Second line of data." @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : str )-> List[Any]: A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt''' A__ = FILE_CONTENT with open(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ ) return filename @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] )-> Optional[Any]: import bza A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.bz2''' A__ = bytes(UpperCamelCase_ , '''utf-8''' ) with bza.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[Any] )-> int: import gzip A__ = str(tmp_path_factory.mktemp('''data''' ) / '''file.txt.gz''' ) A__ = bytes(UpperCamelCase_ , '''utf-8''' ) with gzip.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : str )-> Any: if datasets.config.LZ4_AVAILABLE: import lza.frame A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.lz4''' A__ = bytes(UpperCamelCase_ , '''utf-8''' ) with lza.frame.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : List[Any] , UpperCamelCase_ : Tuple )-> int: if datasets.config.PY7ZR_AVAILABLE: import pyazr A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.7z''' with pyazr.SevenZipFile(UpperCamelCase_ , '''w''' ) as archive: archive.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Tuple , UpperCamelCase_ : int )-> Optional[Any]: import tarfile A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.tar''' with tarfile.TarFile(UpperCamelCase_ , '''w''' ) as f: f.add(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] )-> str: import lzma A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.xz''' A__ = bytes(UpperCamelCase_ , '''utf-8''' ) with lzma.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] , UpperCamelCase_ : List[str] )-> List[str]: import zipfile A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Tuple )-> str: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd A__ = tmp_path_factory.mktemp('''data''' ) / '''file.txt.zst''' A__ = bytes(UpperCamelCase_ , '''utf-8''' ) with zstd.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] )-> int: A__ = tmp_path_factory.mktemp('''data''' ) / '''file.xml''' A__ = textwrap.dedent( '''\ <?xml version="1.0" encoding="UTF-8" ?> <tmx version="1.4"> <header segtype="sentence" srclang="ca" /> <body> <tu> <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv> <tuv xml:lang="en"><seg>Content 1</seg></tuv> </tu> <tu> <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv> <tuv xml:lang="en"><seg>Content 2</seg></tuv> </tu> <tu> <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv> <tuv xml:lang="en"><seg>Content 3</seg></tuv> </tu> <tu> <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv> <tuv xml:lang="en"><seg>Content 4</seg></tuv> </tu> <tu> <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv> <tuv xml:lang="en"><seg>Content 5</seg></tuv> </tu> </body> </tmx>''' ) with open(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ ) return filename _lowercase = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] _lowercase = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] _lowercase = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } _lowercase = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] _lowercase = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( )-> str: return DATA_DICT_OF_LISTS @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : List[Any] )-> List[str]: A__ = datasets.Dataset.from_dict(UpperCamelCase_ ) A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.arrow''' ) dataset.map(cache_file_name=UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict )-> List[str]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.sqlite''' ) with contextlib.closing(sqlitea.connect(UpperCamelCase_ ) ) as con: A__ = con.cursor() cur.execute('''CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)''' ) for item in DATA: cur.execute('''INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)''' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Union[str, Any] )-> Tuple: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.csv''' ) with open(UpperCamelCase_ , '''w''' , newline='''''' ) as f: A__ = csv.DictWriter(UpperCamelCase_ , fieldnames=['''col_1''', '''col_2''', '''col_3'''] ) writer.writeheader() for item in DATA: writer.writerow(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict )-> List[Any]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset2.csv''' ) with open(UpperCamelCase_ , '''w''' , newline='''''' ) as f: A__ = csv.DictWriter(UpperCamelCase_ , fieldnames=['''col_1''', '''col_2''', '''col_3'''] ) writer.writeheader() for item in DATA: writer.writerow(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int , UpperCamelCase_ : Optional[int] )-> List[str]: import bza A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.csv.bz2''' with open(UpperCamelCase_ , '''rb''' ) as f: A__ = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(UpperCamelCase_ , '''wb''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : str , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Optional[int] )-> str: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.csv.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int , UpperCamelCase_ : str , UpperCamelCase_ : Any )-> List[Any]: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.csv.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(csv_path.replace('''.csv''' , '''.CSV''' ) ) ) f.write(UpperCamelCase_ , arcname=os.path.basename(csva_path.replace('''.csv''' , '''.CSV''' ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : str , UpperCamelCase_ : str , UpperCamelCase_ : Optional[Any] )-> Tuple: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset_with_dir.csv.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict )-> Optional[int]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.parquet''' ) A__ = pa.schema( { '''col_1''': pa.string(), '''col_2''': pa.intaa(), '''col_3''': pa.floataa(), } ) with open(UpperCamelCase_ , '''wb''' ) as f: A__ = pq.ParquetWriter(UpperCamelCase_ , schema=UpperCamelCase_ ) A__ = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(UpperCamelCase_ ) )] for k in DATA[0]} , schema=UpperCamelCase_ ) writer.write_table(UpperCamelCase_ ) writer.close() return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] )-> str: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.json''' ) A__ = {'''data''': DATA} with open(UpperCamelCase_ , '''w''' ) as f: json.dump(UpperCamelCase_ , UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] )-> Optional[Any]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.json''' ) A__ = {'''data''': DATA_DICT_OF_LISTS} with open(UpperCamelCase_ , '''w''' ) as f: json.dump(UpperCamelCase_ , UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict )-> List[Any]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.jsonl''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in DATA: f.write(json.dumps(UpperCamelCase_ ) + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : str )-> Tuple: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset2.jsonl''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in DATA: f.write(json.dumps(UpperCamelCase_ ) + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int )-> Optional[int]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset_312.jsonl''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in DATA_312: f.write(json.dumps(UpperCamelCase_ ) + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int )-> Union[str, Any]: A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset-str.jsonl''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in DATA_STR: f.write(json.dumps(UpperCamelCase_ ) + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : str )-> List[str]: import gzip A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.txt.gz''' ) with open(UpperCamelCase_ , '''rb''' ) as orig_file: with gzip.open(UpperCamelCase_ , '''wb''' ) as zipped_file: zipped_file.writelines(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Any , UpperCamelCase_ : Optional[Any] )-> List[Any]: import gzip A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.jsonl.gz''' ) with open(UpperCamelCase_ , '''rb''' ) as orig_file: with gzip.open(UpperCamelCase_ , '''wb''' ) as zipped_file: zipped_file.writelines(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int , UpperCamelCase_ : List[str] , UpperCamelCase_ : List[str] )-> Optional[int]: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.jsonl.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : List[str] , UpperCamelCase_ : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Optional[int] )-> Union[str, Any]: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset_nested.jsonl.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.join('''nested''' , os.path.basename(UpperCamelCase_ ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int , UpperCamelCase_ : List[str] , UpperCamelCase_ : Any )-> Any: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset_with_dir.jsonl.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict , UpperCamelCase_ : Any , UpperCamelCase_ : str )-> Tuple: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.jsonl.tar''' with tarfile.TarFile(UpperCamelCase_ , '''w''' ) as f: f.add(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) f.add(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int , UpperCamelCase_ : List[str] , UpperCamelCase_ : List[str] )-> str: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset_nested.jsonl.tar''' with tarfile.TarFile(UpperCamelCase_ , '''w''' ) as f: f.add(UpperCamelCase_ , arcname=os.path.join('''nested''' , os.path.basename(UpperCamelCase_ ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Any )-> List[str]: A__ = ['''0''', '''1''', '''2''', '''3'''] A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset.txt''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in data: f.write(item + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : int )-> Dict: A__ = ['''0''', '''1''', '''2''', '''3'''] A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset2.txt''' ) with open(UpperCamelCase_ , '''w''' ) as f: for item in data: f.write(item + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Union[str, Any] )-> Union[str, Any]: A__ = ['''0''', '''1''', '''2''', '''3'''] A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.abc''' with open(UpperCamelCase_ , '''w''' ) as f: for item in data: f.write(item + '''\n''' ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : List[str] , UpperCamelCase_ : Any , UpperCamelCase_ : Optional[Any] )-> List[str]: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.text.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Any , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] )-> Tuple: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset_with_dir.text.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) f.write(UpperCamelCase_ , arcname=os.path.join('''main_dir''' , os.path.basename(UpperCamelCase_ ) ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Tuple )-> Any: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.ext.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename('''unsupported.ext''' ) ) f.write(UpperCamelCase_ , arcname=os.path.basename('''unsupported_2.ext''' ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Dict )-> Any: A__ = '''\n'''.join(['''First''', '''Second\u2029with Unicode new line''', '''Third'''] ) A__ = str(tmp_path_factory.mktemp('''data''' ) / '''dataset_with_unicode_new_lines.txt''' ) with open(UpperCamelCase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(UpperCamelCase_ ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( )-> Tuple: return os.path.join('''tests''' , '''features''' , '''data''' , '''test_image_rgb.jpg''' ) @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( )-> Dict: return os.path.join('''tests''' , '''features''' , '''data''' , '''test_audio_44100.wav''' ) @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Union[str, Any] )-> List[Any]: A__ = tmp_path_factory.mktemp('''data''' ) / '''dataset.img.zip''' with zipfile.ZipFile(UpperCamelCase_ , '''w''' ) as f: f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ) ) f.write(UpperCamelCase_ , arcname=os.path.basename(UpperCamelCase_ ).replace('''.jpg''' , '''2.jpg''' ) ) return path @pytest.fixture(scope='''session''' ) def lowerCAmelCase__ ( UpperCamelCase_ : Optional[Any] )-> str: A__ = tmp_path_factory.mktemp('''data_dir''' ) (data_dir / "subdir").mkdir() with open(data_dir / '''subdir''' / '''train.txt''' , '''w''' ) as f: f.write('''foo\n''' * 1_0 ) with open(data_dir / '''subdir''' / '''test.txt''' , '''w''' ) as f: f.write('''bar\n''' * 1_0 ) # hidden file with open(data_dir / '''subdir''' / '''.test.txt''' , '''w''' ) as f: f.write('''bar\n''' * 1_0 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '''.subdir''' / '''train.txt''' , '''w''' ) as f: f.write('''foo\n''' * 1_0 ) with open(data_dir / '''.subdir''' / '''test.txt''' , '''w''' ) as f: f.write('''bar\n''' * 1_0 ) return data_dir
526
0
"""simple docstring""" import math _snake_case = 10 _snake_case = 7 _snake_case = BALLS_PER_COLOUR * NUM_COLOURS def lowerCAmelCase__ ( UpperCamelCase__ = 2_0 ): '''simple docstring''' _a : Dict = math.comb(UpperCamelCase__ , UpperCamelCase__ ) _a : Dict = math.comb(NUM_BALLS - BALLS_PER_COLOUR , UpperCamelCase__ ) _a : List[Any] = NUM_COLOURS * (1 - missing_colour / total) return F"""{result:.9f}""" if __name__ == "__main__": print(solution(20))
389
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _snake_case = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } _snake_case = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } _snake_case = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } _snake_case = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } _snake_case = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } _snake_case = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class UpperCamelCase ( snake_case_ ): UpperCamelCase : Any = VOCAB_FILES_NAMES UpperCamelCase : Dict = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[Any] = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : Any = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION UpperCamelCase : List[Any] = DPRContextEncoderTokenizer class UpperCamelCase ( snake_case_ ): UpperCamelCase : Tuple = VOCAB_FILES_NAMES UpperCamelCase : Optional[Any] = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Dict = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : Optional[int] = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION UpperCamelCase : int = DPRQuestionEncoderTokenizer _snake_case = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) _snake_case = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) _snake_case = r'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(snake_case_ ) class UpperCamelCase : def __call__( self : Any , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Union[bool, str] = False , UpperCAmelCase__ : Union[bool, str] = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : Optional[bool] = None , **UpperCAmelCase__ : Optional[int] , ) -> BatchEncoding: if titles is None and texts is None: return super().__call__( UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , **UpperCAmelCase__ , ) elif titles is None or texts is None: _a : int = titles if texts is None else texts return super().__call__( UpperCAmelCase__ , UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , **UpperCAmelCase__ , ) _a : Optional[int] = titles if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else [titles] _a : Tuple = texts if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else [texts] _a : List[Any] = len(UpperCAmelCase__ ) _a : str = questions if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else [questions] * n_passages assert len(UpperCAmelCase__ ) == len( UpperCAmelCase__ ), f"""There should be as many titles than texts but got {len(UpperCAmelCase__ )} titles and {len(UpperCAmelCase__ )} texts.""" _a : Dict = super().__call__(UpperCAmelCase__ , UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ )["""input_ids"""] _a : int = super().__call__(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ )["""input_ids"""] _a : Optional[int] = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(UpperCAmelCase__ , UpperCAmelCase__ ) ] } if return_attention_mask is not False: _a : int = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) _a : Tuple = attention_mask return self.pad(UpperCAmelCase__ , padding=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors=UpperCAmelCase__ ) def _lowercase ( self : Any , UpperCAmelCase__ : BatchEncoding , UpperCAmelCase__ : DPRReaderOutput , UpperCAmelCase__ : int = 16 , UpperCAmelCase__ : int = 64 , UpperCAmelCase__ : int = 4 , ) -> List[DPRSpanPrediction]: _a : int = reader_input["""input_ids"""] _a , _a , _a : Tuple = reader_output[:3] _a : Optional[Any] = len(UpperCAmelCase__ ) _a : Optional[int] = sorted(range(UpperCAmelCase__ ) , reverse=UpperCAmelCase__ , key=relevance_logits.__getitem__ ) _a : List[DPRReaderOutput] = [] for doc_id in sorted_docs: _a : Tuple = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence _a : Any = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: _a : Optional[Any] = sequence_ids.index(self.pad_token_id ) else: _a : Optional[Any] = len(UpperCAmelCase__ ) _a : Union[str, Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=UpperCAmelCase__ , top_spans=UpperCAmelCase__ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=UpperCAmelCase__ , start_index=UpperCAmelCase__ , end_index=UpperCAmelCase__ , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(UpperCAmelCase__ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def _lowercase ( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , ) -> List[DPRSpanPrediction]: _a : Optional[Any] = [] for start_index, start_score in enumerate(UpperCAmelCase__ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) _a : List[str] = sorted(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : x[1] , reverse=UpperCAmelCase__ ) _a : Optional[Any] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f"""Wrong span indices: [{start_index}:{end_index}]""" _a : List[Any] = end_index - start_index + 1 assert length <= max_answer_length, f"""Span is too long: {length} > {max_answer_length}""" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(UpperCAmelCase__ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(snake_case_ ) class UpperCamelCase ( snake_case_ , snake_case_ ): UpperCamelCase : int = VOCAB_FILES_NAMES UpperCamelCase : int = READER_PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : Dict = READER_PRETRAINED_INIT_CONFIGURATION UpperCamelCase : List[Any] = ['''input_ids''', '''attention_mask'''] UpperCamelCase : Dict = DPRReaderTokenizer
389
1
'''simple docstring''' import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, get_image_size, infer_channel_dimension_format, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_vision_available, logging from ...utils.import_utils import requires_backends if is_vision_available(): import textwrap from PIL import Image, ImageDraw, ImageFont if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: UpperCAmelCase = False UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = 'ybelkada/fonts' def _snake_case ( ) -> Tuple: """simple docstring""" if is_torch_available() and not is_torch_greater_or_equal_than_1_11: raise ImportError( f'You are using torch=={torch.__version__}, but torch>=1.11.0 is required to use ' """Pix2StructImageProcessor. Please upgrade torch.""" ) def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple ) -> List[Any]: """simple docstring""" requires_backends(_SCREAMING_SNAKE_CASE , ["""torch"""] ) _check_torch_version() lowerCAmelCase = image_tensor.unsqueeze(0 ) lowerCAmelCase = torch.nn.functional.unfold(_SCREAMING_SNAKE_CASE , (patch_height, patch_width) , stride=(patch_height, patch_width) ) lowerCAmelCase = patches.reshape(image_tensor.size(0 ) , image_tensor.size(1 ) , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , -1 ) lowerCAmelCase = patches.permute(0 , 4 , 2 , 3 , 1 ).reshape( image_tensor.size(2 ) // patch_height , image_tensor.size(3 ) // patch_width , image_tensor.size(1 ) * patch_height * patch_width , ) return patches.unsqueeze(0 ) def _snake_case ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int = 36 , _SCREAMING_SNAKE_CASE : str = "black" , _SCREAMING_SNAKE_CASE : str = "white" , _SCREAMING_SNAKE_CASE : int = 5 , _SCREAMING_SNAKE_CASE : int = 5 , _SCREAMING_SNAKE_CASE : int = 5 , _SCREAMING_SNAKE_CASE : int = 5 , _SCREAMING_SNAKE_CASE : Optional[bytes] = None , _SCREAMING_SNAKE_CASE : Optional[str] = None , ) -> Image.Image: """simple docstring""" requires_backends(_SCREAMING_SNAKE_CASE , """vision""" ) # Add new lines so that each line is no more than 80 characters. lowerCAmelCase = textwrap.TextWrapper(width=80 ) lowerCAmelCase = wrapper.wrap(text=_SCREAMING_SNAKE_CASE ) lowerCAmelCase = """\n""".join(_SCREAMING_SNAKE_CASE ) if font_bytes is not None and font_path is None: lowerCAmelCase = io.BytesIO(_SCREAMING_SNAKE_CASE ) elif font_path is not None: lowerCAmelCase = font_path else: lowerCAmelCase = hf_hub_download(_SCREAMING_SNAKE_CASE , """Arial.TTF""" ) lowerCAmelCase = ImageFont.truetype(_SCREAMING_SNAKE_CASE , encoding="""UTF-8""" , size=_SCREAMING_SNAKE_CASE ) # Use a temporary canvas to determine the width and height in pixels when # rendering the text. lowerCAmelCase = ImageDraw.Draw(Image.new("""RGB""" , (1, 1) , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase = temp_draw.textbbox((0, 0) , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Create the actual image with a bit of padding around the text. lowerCAmelCase = text_width + left_padding + right_padding lowerCAmelCase = text_height + top_padding + bottom_padding lowerCAmelCase = Image.new("""RGB""" , (image_width, image_height) , _SCREAMING_SNAKE_CASE ) lowerCAmelCase = ImageDraw.Draw(_SCREAMING_SNAKE_CASE ) draw.text(xy=(left_padding, top_padding) , text=_SCREAMING_SNAKE_CASE , fill=_SCREAMING_SNAKE_CASE , font=_SCREAMING_SNAKE_CASE ) return image def _snake_case ( _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : str , **_SCREAMING_SNAKE_CASE : Dict ) -> int: """simple docstring""" requires_backends(_SCREAMING_SNAKE_CASE , """vision""" ) # Convert to PIL image if necessary lowerCAmelCase = to_pil_image(_SCREAMING_SNAKE_CASE ) lowerCAmelCase = render_text(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCAmelCase = max(header_image.width , image.width ) lowerCAmelCase = int(image.height * (new_width / image.width) ) lowerCAmelCase = int(header_image.height * (new_width / header_image.width) ) lowerCAmelCase = Image.new("""RGB""" , (new_width, new_height + new_header_height) , """white""" ) new_image.paste(header_image.resize((new_width, new_header_height) ) , (0, 0) ) new_image.paste(image.resize((new_width, new_height) ) , (0, new_header_height) ) # Convert back to the original framework if necessary lowerCAmelCase = to_numpy_array(_SCREAMING_SNAKE_CASE ) if infer_channel_dimension_format(_SCREAMING_SNAKE_CASE ) == ChannelDimension.LAST: lowerCAmelCase = to_channel_dimension_format(_SCREAMING_SNAKE_CASE , ChannelDimension.LAST ) return new_image class __snake_case( _lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : List[Any] = ["flattened_patches"] def __init__( self , A_ = True , A_ = True , A_ = None , A_ = 2048 , A_ = False , **A_ , ) -> None: super().__init__(**A_ ) lowerCAmelCase = patch_size if patch_size is not None else {"""height""": 16, """width""": 16} lowerCAmelCase = do_normalize lowerCAmelCase = do_convert_rgb lowerCAmelCase = max_patches lowerCAmelCase = is_vqa def __snake_case ( self , A_ , A_ , A_ , **A_ ) -> np.ndarray: requires_backends(self.extract_flattened_patches , """torch""" ) _check_torch_version() # convert to torch lowerCAmelCase = to_channel_dimension_format(A_ , ChannelDimension.FIRST ) lowerCAmelCase = torch.from_numpy(A_ ) lowerCAmelCase, lowerCAmelCase = patch_size["""height"""], patch_size["""width"""] lowerCAmelCase, lowerCAmelCase = get_image_size(A_ ) # maximize scale s.t. lowerCAmelCase = math.sqrt(max_patches * (patch_height / image_height) * (patch_width / image_width) ) lowerCAmelCase = max(min(math.floor(scale * image_height / patch_height ) , A_ ) , 1 ) lowerCAmelCase = max(min(math.floor(scale * image_width / patch_width ) , A_ ) , 1 ) lowerCAmelCase = max(num_feasible_rows * patch_height , 1 ) lowerCAmelCase = max(num_feasible_cols * patch_width , 1 ) lowerCAmelCase = torch.nn.functional.interpolate( image.unsqueeze(0 ) , size=(resized_height, resized_width) , mode="""bilinear""" , align_corners=A_ , antialias=A_ , ).squeeze(0 ) # [1, rows, columns, patch_height * patch_width * image_channels] lowerCAmelCase = torch_extract_patches(A_ , A_ , A_ ) lowerCAmelCase = patches.shape lowerCAmelCase = patches_shape[1] lowerCAmelCase = patches_shape[2] lowerCAmelCase = patches_shape[3] # [rows * columns, patch_height * patch_width * image_channels] lowerCAmelCase = patches.reshape([rows * columns, depth] ) # [rows * columns, 1] lowerCAmelCase = torch.arange(A_ ).reshape([rows, 1] ).repeat(1 , A_ ).reshape([rows * columns, 1] ) lowerCAmelCase = torch.arange(A_ ).reshape([1, columns] ).repeat(A_ , 1 ).reshape([rows * columns, 1] ) # Offset by 1 so the ids do not contain zeros, which represent padding. row_ids += 1 col_ids += 1 # Prepare additional patch features. # [rows * columns, 1] lowerCAmelCase = row_ids.to(torch.floataa ) lowerCAmelCase = col_ids.to(torch.floataa ) # [rows * columns, 2 + patch_height * patch_width * image_channels] lowerCAmelCase = torch.cat([row_ids, col_ids, patches] , -1 ) # [max_patches, 2 + patch_height * patch_width * image_channels] lowerCAmelCase = torch.nn.functional.pad(A_ , [0, 0, 0, max_patches - (rows * columns)] ).float() lowerCAmelCase = to_numpy_array(A_ ) return result def __snake_case ( self , A_ , A_ = None , **A_ ) -> np.ndarray: if image.dtype == np.uinta: lowerCAmelCase = image.astype(np.floataa ) # take mean across the whole `image` lowerCAmelCase = np.mean(A_ ) lowerCAmelCase = np.std(A_ ) lowerCAmelCase = max(A_ , 1.0 / math.sqrt(np.prod(image.shape ) ) ) return normalize(A_ , mean=A_ , std=A_ , **A_ ) def __snake_case ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) -> ImageInput: lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCAmelCase = patch_size if patch_size is not None else self.patch_size lowerCAmelCase = max_patches if max_patches is not None else self.max_patches lowerCAmelCase = self.is_vqa if kwargs.get("""data_format""" , A_ ) is not None: raise ValueError("""data_format is not an accepted input as the outputs are """ ) lowerCAmelCase = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCAmelCase = [convert_to_rgb(A_ ) for image in images] # All transformations expect numpy arrays. lowerCAmelCase = [to_numpy_array(A_ ) for image in images] if is_vqa: if header_text is None: raise ValueError("""A header text must be provided for VQA models.""" ) lowerCAmelCase = kwargs.pop("""font_bytes""" , A_ ) lowerCAmelCase = kwargs.pop("""font_path""" , A_ ) if isinstance(A_ , A_ ): lowerCAmelCase = [header_text] * len(A_ ) lowerCAmelCase = [ render_header(A_ , header_text[i] , font_bytes=A_ , font_path=A_ ) for i, image in enumerate(A_ ) ] if do_normalize: lowerCAmelCase = [self.normalize(image=A_ ) for image in images] # convert to torch tensor and permute lowerCAmelCase = [ self.extract_flattened_patches(image=A_ , max_patches=A_ , patch_size=A_ ) for image in images ] # create attention mask in numpy lowerCAmelCase = [(image.sum(axis=-1 ) != 0).astype(np.floataa ) for image in images] lowerCAmelCase = BatchFeature( data={"""flattened_patches""": images, """attention_mask""": attention_masks} , tensor_type=A_ ) return encoded_outputs
344
'''simple docstring''' import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class __snake_case( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ) -> Tuple: lowerCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCAmelCase = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(A_ ) lowerCAmelCase = -1 lowerCAmelCase = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(A_ ) lowerCAmelCase = model.generate(A_ , max_new_tokens=10 , do_sample=A_ ) lowerCAmelCase = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: lowerCAmelCase = TextStreamer(A_ ) model.generate(A_ , max_new_tokens=10 , do_sample=A_ , streamer=A_ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCAmelCase = cs.out[:-1] self.assertEqual(A_ , A_ ) def __snake_case ( self ) -> List[Any]: lowerCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCAmelCase = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(A_ ) lowerCAmelCase = -1 lowerCAmelCase = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(A_ ) lowerCAmelCase = model.generate(A_ , max_new_tokens=10 , do_sample=A_ ) lowerCAmelCase = tokenizer.decode(greedy_ids[0] ) lowerCAmelCase = TextIteratorStreamer(A_ ) lowerCAmelCase = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} lowerCAmelCase = Thread(target=model.generate , kwargs=A_ ) thread.start() lowerCAmelCase = """""" for new_text in streamer: streamer_text += new_text self.assertEqual(A_ , A_ ) def __snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCAmelCase = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(A_ ) lowerCAmelCase = -1 lowerCAmelCase = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(A_ ) lowerCAmelCase = model.generate(A_ , max_new_tokens=10 , do_sample=A_ ) lowerCAmelCase = greedy_ids[:, input_ids.shape[1] :] lowerCAmelCase = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: lowerCAmelCase = TextStreamer(A_ , skip_prompt=A_ ) model.generate(A_ , max_new_tokens=10 , do_sample=A_ , streamer=A_ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCAmelCase = cs.out[:-1] self.assertEqual(A_ , A_ ) def __snake_case ( self ) -> int: # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them lowerCAmelCase = AutoTokenizer.from_pretrained("""distilgpt2""" ) lowerCAmelCase = AutoModelForCausalLM.from_pretrained("""distilgpt2""" ).to(A_ ) lowerCAmelCase = -1 lowerCAmelCase = torch.ones((1, 5) , device=A_ ).long() * model.config.bos_token_id with CaptureStdout() as cs: lowerCAmelCase = TextStreamer(A_ , skip_special_tokens=A_ ) model.generate(A_ , max_new_tokens=1 , do_sample=A_ , streamer=A_ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token lowerCAmelCase = cs.out[:-1] # Remove the final "\n" lowerCAmelCase = tokenizer(A_ , return_tensors="""pt""" ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def __snake_case ( self ) -> Tuple: lowerCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCAmelCase = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(A_ ) lowerCAmelCase = -1 lowerCAmelCase = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(A_ ) lowerCAmelCase = TextIteratorStreamer(A_ , timeout=0.0_0_1 ) lowerCAmelCase = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} lowerCAmelCase = Thread(target=model.generate , kwargs=A_ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(A_ ): lowerCAmelCase = """""" for new_text in streamer: streamer_text += new_text
344
1
"""simple docstring""" from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split __A = datasets.load_iris() __A = np.array(data["""data"""]) __A = np.array(data["""target"""]) __A = data["""target_names"""] __A , __A , __A , __A = train_test_split(X, y) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" return np.linalg.norm(np.array(_SCREAMING_SNAKE_CASE ) - np.array(_SCREAMING_SNAKE_CASE ) ) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=5 ) ->Optional[Any]: """simple docstring""" lowerCAmelCase__ :str = zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # List of distances of all points from the point to be classified lowerCAmelCase__ :Optional[Any] = [] for data_point in data: lowerCAmelCase__ :Tuple = euclidean_distance(data_point[0] , _SCREAMING_SNAKE_CASE ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. lowerCAmelCase__ :str = [i[1] for i in sorted(_SCREAMING_SNAKE_CASE )[:k]] # Most commonly occurring class among them # is the class into which the point is classified lowerCAmelCase__ :Optional[int] = Counter(_SCREAMING_SNAKE_CASE ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
93
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters A_ = (7_20, 12_80) # Height, Width A_ = (0.4, 0.6) # if height or width lower than this scale, drop it. A_ = 1 / 1_00 A_ = '''''' A_ = '''''' A_ = '''''' A_ = 2_50 def UpperCAmelCase__ (): """simple docstring""" _snake_case , _snake_case : Any = get_dataset(snake_case__ , snake_case__ ) for index in range(snake_case__ ): _snake_case : List[Any] = random.sample(range(len(snake_case__ ) ) , 4 ) _snake_case , _snake_case , _snake_case : Tuple = update_image_and_anno( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , filter_scale=snake_case__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _snake_case : List[Any] = random_chars(32 ) _snake_case : List[Any] = path.split(os.sep )[-1].rsplit(""".""" , 1 )[0] _snake_case : Union[str, Any] = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , snake_case__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) _snake_case : Optional[Any] = [] for anno in new_annos: _snake_case : List[str] = anno[3] - anno[1] _snake_case : Any = anno[4] - anno[2] _snake_case : Any = anno[1] + width / 2 _snake_case : List[Any] = anno[2] + height / 2 _snake_case : Any = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(snake_case__ ) with open(F"{file_root}.txt" , """w""" ) as outfile: outfile.write("""\n""".join(line for line in annos_list ) ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[int] = [] for label_file in glob.glob(os.path.join(snake_case__ , """*.txt""" ) ): _snake_case : Optional[Any] = label_file.split(os.sep )[-1].rsplit(""".""" , 1 )[0] with open(snake_case__ ) as in_file: _snake_case : Union[str, Any] = in_file.readlines() _snake_case : Optional[Any] = os.path.join(snake_case__ , F"{label_name}.jpg" ) _snake_case : Tuple = [] for obj_list in obj_lists: _snake_case : Optional[int] = obj_list.rstrip("""\n""" ).split(""" """ ) _snake_case : Union[str, Any] = float(obj[1] ) - float(obj[3] ) / 2 _snake_case : Tuple = float(obj[2] ) - float(obj[4] ) / 2 _snake_case : List[str] = float(obj[1] ) + float(obj[3] ) / 2 _snake_case : Union[str, Any] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(snake_case__ ) labels.append(snake_case__ ) return img_paths, labels def UpperCAmelCase__ (snake_case__ : list , snake_case__ : list , snake_case__ : list[int] , snake_case__ : tuple[int, int] , snake_case__ : tuple[float, float] , snake_case__ : float = 0.0 , ): """simple docstring""" _snake_case : str = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) _snake_case : Any = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _snake_case : List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _snake_case : List[str] = int(scale_x * output_size[1] ) _snake_case : Any = int(scale_y * output_size[0] ) _snake_case : Optional[Any] = [] _snake_case : List[str] = [] for i, index in enumerate(snake_case__ ): _snake_case : str = all_img_list[index] path_list.append(snake_case__ ) _snake_case : Any = all_annos[index] _snake_case : Tuple = cva.imread(snake_case__ ) if i == 0: # top-left _snake_case : Tuple = cva.resize(snake_case__ , (divid_point_x, divid_point_y) ) _snake_case : int = img for bbox in img_annos: _snake_case : str = bbox[1] * scale_x _snake_case : Optional[int] = bbox[2] * scale_y _snake_case : Dict = bbox[3] * scale_x _snake_case : Any = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _snake_case : Dict = cva.resize(snake_case__ , (output_size[1] - divid_point_x, divid_point_y) ) _snake_case : int = img for bbox in img_annos: _snake_case : Any = scale_x + bbox[1] * (1 - scale_x) _snake_case : Union[str, Any] = bbox[2] * scale_y _snake_case : List[str] = scale_x + bbox[3] * (1 - scale_x) _snake_case : Tuple = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _snake_case : Union[str, Any] = cva.resize(snake_case__ , (divid_point_x, output_size[0] - divid_point_y) ) _snake_case : Optional[Any] = img for bbox in img_annos: _snake_case : int = bbox[1] * scale_x _snake_case : Union[str, Any] = scale_y + bbox[2] * (1 - scale_y) _snake_case : Optional[int] = bbox[3] * scale_x _snake_case : Any = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _snake_case : int = cva.resize( snake_case__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _snake_case : List[Any] = img for bbox in img_annos: _snake_case : str = scale_x + bbox[1] * (1 - scale_x) _snake_case : Any = scale_y + bbox[2] * (1 - scale_y) _snake_case : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) _snake_case : Union[str, Any] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _snake_case : Optional[int] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" _snake_case : int = ascii_lowercase + digits return "".join(random.choice(snake_case__ ) for _ in range(snake_case__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
609
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool _lowercase : Optional[Any] ={ '''Acehnese Arabic''': '''ace_Arab''', '''Acehnese Latin''': '''ace_Latn''', '''Mesopotamian Arabic''': '''acm_Arab''', '''Ta\'izzi-Adeni Arabic''': '''acq_Arab''', '''Tunisian Arabic''': '''aeb_Arab''', '''Afrikaans''': '''afr_Latn''', '''South Levantine Arabic''': '''ajp_Arab''', '''Akan''': '''aka_Latn''', '''Amharic''': '''amh_Ethi''', '''North Levantine Arabic''': '''apc_Arab''', '''Modern Standard Arabic''': '''arb_Arab''', '''Modern Standard Arabic Romanized''': '''arb_Latn''', '''Najdi Arabic''': '''ars_Arab''', '''Moroccan Arabic''': '''ary_Arab''', '''Egyptian Arabic''': '''arz_Arab''', '''Assamese''': '''asm_Beng''', '''Asturian''': '''ast_Latn''', '''Awadhi''': '''awa_Deva''', '''Central Aymara''': '''ayr_Latn''', '''South Azerbaijani''': '''azb_Arab''', '''North Azerbaijani''': '''azj_Latn''', '''Bashkir''': '''bak_Cyrl''', '''Bambara''': '''bam_Latn''', '''Balinese''': '''ban_Latn''', '''Belarusian''': '''bel_Cyrl''', '''Bemba''': '''bem_Latn''', '''Bengali''': '''ben_Beng''', '''Bhojpuri''': '''bho_Deva''', '''Banjar Arabic''': '''bjn_Arab''', '''Banjar Latin''': '''bjn_Latn''', '''Standard Tibetan''': '''bod_Tibt''', '''Bosnian''': '''bos_Latn''', '''Buginese''': '''bug_Latn''', '''Bulgarian''': '''bul_Cyrl''', '''Catalan''': '''cat_Latn''', '''Cebuano''': '''ceb_Latn''', '''Czech''': '''ces_Latn''', '''Chokwe''': '''cjk_Latn''', '''Central Kurdish''': '''ckb_Arab''', '''Crimean Tatar''': '''crh_Latn''', '''Welsh''': '''cym_Latn''', '''Danish''': '''dan_Latn''', '''German''': '''deu_Latn''', '''Southwestern Dinka''': '''dik_Latn''', '''Dyula''': '''dyu_Latn''', '''Dzongkha''': '''dzo_Tibt''', '''Greek''': '''ell_Grek''', '''English''': '''eng_Latn''', '''Esperanto''': '''epo_Latn''', '''Estonian''': '''est_Latn''', '''Basque''': '''eus_Latn''', '''Ewe''': '''ewe_Latn''', '''Faroese''': '''fao_Latn''', '''Fijian''': '''fij_Latn''', '''Finnish''': '''fin_Latn''', '''Fon''': '''fon_Latn''', '''French''': '''fra_Latn''', '''Friulian''': '''fur_Latn''', '''Nigerian Fulfulde''': '''fuv_Latn''', '''Scottish Gaelic''': '''gla_Latn''', '''Irish''': '''gle_Latn''', '''Galician''': '''glg_Latn''', '''Guarani''': '''grn_Latn''', '''Gujarati''': '''guj_Gujr''', '''Haitian Creole''': '''hat_Latn''', '''Hausa''': '''hau_Latn''', '''Hebrew''': '''heb_Hebr''', '''Hindi''': '''hin_Deva''', '''Chhattisgarhi''': '''hne_Deva''', '''Croatian''': '''hrv_Latn''', '''Hungarian''': '''hun_Latn''', '''Armenian''': '''hye_Armn''', '''Igbo''': '''ibo_Latn''', '''Ilocano''': '''ilo_Latn''', '''Indonesian''': '''ind_Latn''', '''Icelandic''': '''isl_Latn''', '''Italian''': '''ita_Latn''', '''Javanese''': '''jav_Latn''', '''Japanese''': '''jpn_Jpan''', '''Kabyle''': '''kab_Latn''', '''Jingpho''': '''kac_Latn''', '''Kamba''': '''kam_Latn''', '''Kannada''': '''kan_Knda''', '''Kashmiri Arabic''': '''kas_Arab''', '''Kashmiri Devanagari''': '''kas_Deva''', '''Georgian''': '''kat_Geor''', '''Central Kanuri Arabic''': '''knc_Arab''', '''Central Kanuri Latin''': '''knc_Latn''', '''Kazakh''': '''kaz_Cyrl''', '''Kabiyè''': '''kbp_Latn''', '''Kabuverdianu''': '''kea_Latn''', '''Khmer''': '''khm_Khmr''', '''Kikuyu''': '''kik_Latn''', '''Kinyarwanda''': '''kin_Latn''', '''Kyrgyz''': '''kir_Cyrl''', '''Kimbundu''': '''kmb_Latn''', '''Northern Kurdish''': '''kmr_Latn''', '''Kikongo''': '''kon_Latn''', '''Korean''': '''kor_Hang''', '''Lao''': '''lao_Laoo''', '''Ligurian''': '''lij_Latn''', '''Limburgish''': '''lim_Latn''', '''Lingala''': '''lin_Latn''', '''Lithuanian''': '''lit_Latn''', '''Lombard''': '''lmo_Latn''', '''Latgalian''': '''ltg_Latn''', '''Luxembourgish''': '''ltz_Latn''', '''Luba-Kasai''': '''lua_Latn''', '''Ganda''': '''lug_Latn''', '''Luo''': '''luo_Latn''', '''Mizo''': '''lus_Latn''', '''Standard Latvian''': '''lvs_Latn''', '''Magahi''': '''mag_Deva''', '''Maithili''': '''mai_Deva''', '''Malayalam''': '''mal_Mlym''', '''Marathi''': '''mar_Deva''', '''Minangkabau Arabic ''': '''min_Arab''', '''Minangkabau Latin''': '''min_Latn''', '''Macedonian''': '''mkd_Cyrl''', '''Plateau Malagasy''': '''plt_Latn''', '''Maltese''': '''mlt_Latn''', '''Meitei Bengali''': '''mni_Beng''', '''Halh Mongolian''': '''khk_Cyrl''', '''Mossi''': '''mos_Latn''', '''Maori''': '''mri_Latn''', '''Burmese''': '''mya_Mymr''', '''Dutch''': '''nld_Latn''', '''Norwegian Nynorsk''': '''nno_Latn''', '''Norwegian Bokmål''': '''nob_Latn''', '''Nepali''': '''npi_Deva''', '''Northern Sotho''': '''nso_Latn''', '''Nuer''': '''nus_Latn''', '''Nyanja''': '''nya_Latn''', '''Occitan''': '''oci_Latn''', '''West Central Oromo''': '''gaz_Latn''', '''Odia''': '''ory_Orya''', '''Pangasinan''': '''pag_Latn''', '''Eastern Panjabi''': '''pan_Guru''', '''Papiamento''': '''pap_Latn''', '''Western Persian''': '''pes_Arab''', '''Polish''': '''pol_Latn''', '''Portuguese''': '''por_Latn''', '''Dari''': '''prs_Arab''', '''Southern Pashto''': '''pbt_Arab''', '''Ayacucho Quechua''': '''quy_Latn''', '''Romanian''': '''ron_Latn''', '''Rundi''': '''run_Latn''', '''Russian''': '''rus_Cyrl''', '''Sango''': '''sag_Latn''', '''Sanskrit''': '''san_Deva''', '''Santali''': '''sat_Olck''', '''Sicilian''': '''scn_Latn''', '''Shan''': '''shn_Mymr''', '''Sinhala''': '''sin_Sinh''', '''Slovak''': '''slk_Latn''', '''Slovenian''': '''slv_Latn''', '''Samoan''': '''smo_Latn''', '''Shona''': '''sna_Latn''', '''Sindhi''': '''snd_Arab''', '''Somali''': '''som_Latn''', '''Southern Sotho''': '''sot_Latn''', '''Spanish''': '''spa_Latn''', '''Tosk Albanian''': '''als_Latn''', '''Sardinian''': '''srd_Latn''', '''Serbian''': '''srp_Cyrl''', '''Swati''': '''ssw_Latn''', '''Sundanese''': '''sun_Latn''', '''Swedish''': '''swe_Latn''', '''Swahili''': '''swh_Latn''', '''Silesian''': '''szl_Latn''', '''Tamil''': '''tam_Taml''', '''Tatar''': '''tat_Cyrl''', '''Telugu''': '''tel_Telu''', '''Tajik''': '''tgk_Cyrl''', '''Tagalog''': '''tgl_Latn''', '''Thai''': '''tha_Thai''', '''Tigrinya''': '''tir_Ethi''', '''Tamasheq Latin''': '''taq_Latn''', '''Tamasheq Tifinagh''': '''taq_Tfng''', '''Tok Pisin''': '''tpi_Latn''', '''Tswana''': '''tsn_Latn''', '''Tsonga''': '''tso_Latn''', '''Turkmen''': '''tuk_Latn''', '''Tumbuka''': '''tum_Latn''', '''Turkish''': '''tur_Latn''', '''Twi''': '''twi_Latn''', '''Central Atlas Tamazight''': '''tzm_Tfng''', '''Uyghur''': '''uig_Arab''', '''Ukrainian''': '''ukr_Cyrl''', '''Umbundu''': '''umb_Latn''', '''Urdu''': '''urd_Arab''', '''Northern Uzbek''': '''uzn_Latn''', '''Venetian''': '''vec_Latn''', '''Vietnamese''': '''vie_Latn''', '''Waray''': '''war_Latn''', '''Wolof''': '''wol_Latn''', '''Xhosa''': '''xho_Latn''', '''Eastern Yiddish''': '''ydd_Hebr''', '''Yoruba''': '''yor_Latn''', '''Yue Chinese''': '''yue_Hant''', '''Chinese Simplified''': '''zho_Hans''', '''Chinese Traditional''': '''zho_Hant''', '''Standard Malay''': '''zsm_Latn''', '''Zulu''': '''zul_Latn''', } class SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_ ): '''simple docstring''' lowercase : Dict = "facebook/nllb-200-distilled-600M" lowercase : Optional[Any] = ( "This is a tool that translates text from a language to another. It takes three inputs: `text`, which should " "be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, " "which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in " "plain English, such as 'Romanian', or 'Albanian'. It returns the text translated in `tgt_lang`." ) lowercase : List[str] = "translator" lowercase : List[str] = AutoTokenizer lowercase : int = AutoModelForSeqaSeqLM lowercase : Dict = LANGUAGE_CODES lowercase : Dict = ["text", "text", "text"] lowercase : List[Any] = ["text"] def SCREAMING_SNAKE_CASE_ ( self : str , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Any: if src_lang not in self.lang_to_code: raise ValueError(f'{src_lang} is not a supported language.' ) if tgt_lang not in self.lang_to_code: raise ValueError(f'{tgt_lang} is not a supported language.' ) A : str =self.lang_to_code[src_lang] A : Dict =self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( SCREAMING_SNAKE_CASE__ , return_tensors='pt' , src_lang=SCREAMING_SNAKE_CASE__ , tgt_lang=SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> int: return self.model.generate(**SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Any: return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=SCREAMING_SNAKE_CASE__ )
661
import math def A__ ( lowercase: int ) -> list: A : Optional[Any] =[True] * n A : Tuple =False A : List[Any] =False A : Dict =True for i in range(3, int(n**0.5 + 1 ), 2 ): A : Dict =i * 2 while index < n: A : Dict =False A : Dict =index + i A : Tuple =[2] for i in range(3, lowercase, 2 ): if is_prime[i]: primes.append(lowercase ) return primes def A__ ( lowercase: int = 999_966_663_333 ) -> int: A : Optional[int] =math.floor(math.sqrt(lowercase ) ) + 100 A : Optional[int] =prime_sieve(lowercase ) A : Optional[Any] =0 A : List[Any] =0 A : Union[str, Any] =primes[prime_index] while (last_prime**2) <= limit: A : Tuple =primes[prime_index + 1] A : Optional[int] =last_prime**2 A : Tuple =next_prime**2 # Get numbers divisible by lps(current) A : int =lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) A : List[Any] =upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps A : Any =0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair A : List[str] =next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
661
1
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) lowerCamelCase__ = logging.getLogger(__name__) def _lowerCamelCase( ) -> Any: __snake_case = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path" , type=__snake_case , default="data/dump.txt" , help="The path to the data." ) parser.add_argument("--tokenizer_type" , type=__snake_case , default="bert" , choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name" , type=__snake_case , default="bert-base-uncased" , help="The tokenizer to use." ) parser.add_argument("--dump_file" , type=__snake_case , default="data/dump" , help="The dump file prefix." ) __snake_case = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": __snake_case = BertTokenizer.from_pretrained(args.tokenizer_name ) __snake_case = tokenizer.special_tokens_map["cls_token"] # `[CLS]` __snake_case = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": __snake_case = RobertaTokenizer.from_pretrained(args.tokenizer_name ) __snake_case = tokenizer.special_tokens_map["cls_token"] # `<s>` __snake_case = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": __snake_case = GPTaTokenizer.from_pretrained(args.tokenizer_name ) __snake_case = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` __snake_case = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path , "r" , encoding="utf8" ) as fp: __snake_case = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__snake_case )} examples to process.""" ) __snake_case = [] __snake_case = 0 __snake_case = 1_0000 __snake_case = time.time() for text in data: __snake_case = f"""{bos} {text.strip()} {sep}""" __snake_case = tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) rslt.append(__snake_case ) iter += 1 if iter % interval == 0: __snake_case = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) __snake_case = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__snake_case )} examples processed.""" ) __snake_case = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" __snake_case = tokenizer.vocab_size if vocab_size < (1 << 16): __snake_case = [np.uintaa(__snake_case ) for d in rslt] else: __snake_case = [np.intaa(__snake_case ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__snake_case , "wb" ) as handle: pickle.dump(rslt_ , __snake_case , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
524
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): """simple docstring""" __snake_case = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small" ) __snake_case = AutoTokenizer.from_pretrained("google/mt5-small" ) __snake_case = tokenizer("Hello there" ,return_tensors="np" ).input_ids __snake_case = tokenizer("Hi I am" ,return_tensors="np" ).input_ids __snake_case = shift_tokens_right(_lowerCAmelCase ,model.config.pad_token_id ,model.config.decoder_start_token_id ) __snake_case = model(_lowerCAmelCase ,decoder_input_ids=_lowerCAmelCase ).logits __snake_case = optax.softmax_cross_entropy(_lowerCAmelCase ,onehot(_lowerCAmelCase ,logits.shape[-1] ) ).mean() __snake_case = -(labels.shape[-1] * loss.item()) __snake_case = -8_4.9_1_2_7 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
524
1
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCAmelCase__ = re.compile(r'\s+') def __lowercase ( _UpperCAmelCase ) -> Dict: '''simple docstring''' return {"hash": hashlib.mda(re.sub(UpperCamelCase__ , "" , example["content"] ).encode("utf-8" ) ).hexdigest()} def __lowercase ( _UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' __lowercase = [len(UpperCamelCase__ ) for line in example["content"].splitlines()] return {"line_mean": np.mean(UpperCamelCase__ ), "line_max": max(UpperCamelCase__ )} def __lowercase ( _UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' __lowercase = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __lowercase ( _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: '''simple docstring''' if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __lowercase ( _UpperCAmelCase , _UpperCAmelCase=5 ) -> List[Any]: '''simple docstring''' __lowercase = ["auto-generated", "autogenerated", "automatically generated"] __lowercase = example["content"].splitlines() for _, line in zip(range(UpperCamelCase__ ) , UpperCamelCase__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __lowercase ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ) -> int: '''simple docstring''' __lowercase = ["unit tests", "test file", "configuration file"] __lowercase = example["content"].splitlines() __lowercase = 0 __lowercase = 0 # first test for _, line in zip(range(UpperCamelCase__ ) , UpperCamelCase__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test __lowercase = example["content"].count("\n" ) __lowercase = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __lowercase ( _UpperCAmelCase ) -> Dict: '''simple docstring''' __lowercase = ["def ", "class ", "for ", "while "] __lowercase = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __lowercase ( _UpperCAmelCase , _UpperCAmelCase=4 ) -> List[str]: '''simple docstring''' __lowercase = example["content"].splitlines() __lowercase = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __lowercase ( _UpperCAmelCase ) -> Dict: '''simple docstring''' __lowercase = tokenizer(example["content"] , truncation=UpperCamelCase__ )["input_ids"] __lowercase = len(example["content"] ) / len(UpperCamelCase__ ) return {"ratio": ratio} def __lowercase ( _UpperCAmelCase ) -> str: '''simple docstring''' __lowercase = {} results.update(get_hash(UpperCamelCase__ ) ) results.update(line_stats(UpperCamelCase__ ) ) results.update(alpha_stats(UpperCamelCase__ ) ) results.update(char_token_ratio(UpperCamelCase__ ) ) results.update(is_autogenerated(UpperCamelCase__ ) ) results.update(is_config_or_test(UpperCamelCase__ ) ) results.update(has_no_keywords(UpperCamelCase__ ) ) results.update(has_few_assignments(UpperCamelCase__ ) ) return results def __lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: '''simple docstring''' if not check_uniques(UpperCamelCase__ , UpperCamelCase__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __lowercase ( _UpperCAmelCase ) -> int: '''simple docstring''' with open(UpperCamelCase__ , "rb" ) as f_in: with gzip.open(str(UpperCamelCase__ ) + ".gz" , "wb" , compresslevel=6 ) as f_out: shutil.copyfileobj(UpperCamelCase__ , UpperCamelCase__ ) os.unlink(UpperCamelCase__ ) # Settings lowerCAmelCase__ = HfArgumentParser(PreprocessingArguments) lowerCAmelCase__ = parser.parse_args() if args.num_workers is None: lowerCAmelCase__ = multiprocessing.cpu_count() lowerCAmelCase__ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCAmelCase__ = time.time() lowerCAmelCase__ = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCAmelCase__ = time.time() lowerCAmelCase__ = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCAmelCase__ = set(ds.unique('hash')) lowerCAmelCase__ = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCAmelCase__ = time.time() lowerCAmelCase__ = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCAmelCase__ = time.time() lowerCAmelCase__ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCAmelCase__ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCAmelCase__ = output_dir / """data""" data_dir.mkdir(exist_ok=True) lowerCAmelCase__ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCAmelCase__ = str(data_dir / F"file-{file_number+1:012}.json") lowerCAmelCase__ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
720
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def __lowercase ( _UpperCAmelCase ) -> int: '''simple docstring''' __lowercase = SwinConfig(image_size=192 ) if "base" in model_name: __lowercase = 6 __lowercase = 128 __lowercase = (2, 2, 18, 2) __lowercase = (4, 8, 16, 32) elif "large" in model_name: __lowercase = 12 __lowercase = 192 __lowercase = (2, 2, 18, 2) __lowercase = (6, 12, 24, 48) else: raise ValueError("Model not supported, only supports base and large variants" ) __lowercase = window_size __lowercase = embed_dim __lowercase = depths __lowercase = num_heads return config def __lowercase ( _UpperCAmelCase ) -> List[str]: '''simple docstring''' if "encoder.mask_token" in name: __lowercase = name.replace("encoder.mask_token" , "embeddings.mask_token" ) if "encoder.patch_embed.proj" in name: __lowercase = name.replace("encoder.patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "encoder.patch_embed.norm" in name: __lowercase = name.replace("encoder.patch_embed.norm" , "embeddings.norm" ) if "attn.proj" in name: __lowercase = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: __lowercase = name.replace("attn" , "attention.self" ) if "norm1" in name: __lowercase = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: __lowercase = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: __lowercase = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __lowercase = name.replace("mlp.fc2" , "output.dense" ) if name == "encoder.norm.weight": __lowercase = "layernorm.weight" if name == "encoder.norm.bias": __lowercase = "layernorm.bias" if "decoder" in name: pass else: __lowercase = "swin." + name return name def __lowercase ( _UpperCAmelCase , _UpperCAmelCase ) -> Any: '''simple docstring''' for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(_UpperCAmelCase ) if "attn_mask" in key: pass elif "qkv" in key: __lowercase = key.split("." ) __lowercase = int(key_split[2] ) __lowercase = int(key_split[4] ) __lowercase = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[ dim : dim * 2, : ] __lowercase = val[-dim:, :] else: __lowercase = val[ :dim ] __lowercase = val[ dim : dim * 2 ] __lowercase = val[ -dim: ] else: __lowercase = val return orig_state_dict def __lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: '''simple docstring''' __lowercase = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] __lowercase = get_swin_config(_UpperCAmelCase ) __lowercase = SwinForMaskedImageModeling(_UpperCAmelCase ) model.eval() __lowercase = convert_state_dict(_UpperCAmelCase , _UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) __lowercase = "http://images.cocodataset.org/val2017/000000039769.jpg" __lowercase = ViTImageProcessor(size={"height": 192, "width": 192} ) __lowercase = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) __lowercase = image_processor(images=_UpperCAmelCase , return_tensors="pt" ) with torch.no_grad(): __lowercase = model(**_UpperCAmelCase ).logits print(outputs.keys() ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_UpperCAmelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f'''Pushing model and image processor for {model_name} to hub''' ) model.push_to_hub(f'''microsoft/{model_name}''' ) image_processor.push_to_hub(f'''microsoft/{model_name}''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='swin-base-simmim-window6-192', type=str, choices=['swin-base-simmim-window6-192', 'swin-large-simmim-window12-192'], help='Name of the Swin SimMIM model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default='/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth', type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) lowerCAmelCase__ = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
576
0
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase : def __init__( self : Union[str, Any] , snake_case__ : Dict , snake_case__ : Tuple=1_2 , snake_case__ : Dict=7 , snake_case__ : Union[str, Any]=True , snake_case__ : str=True , snake_case__ : List[str]=True , snake_case__ : Dict=9_9 , snake_case__ : Optional[Any]=3_2 , snake_case__ : Union[str, Any]=3_2 , snake_case__ : Dict=2 , snake_case__ : Any=4 , snake_case__ : Union[str, Any]=3_7 , snake_case__ : Dict=0.1 , snake_case__ : int=0.1 , snake_case__ : List[str]=5_1_2 , snake_case__ : Dict=0.02 , snake_case__ : List[str]=0 , snake_case__ : Any=None , ): """simple docstring""" SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = seq_length SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_input_mask SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = projection_dim SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = scope SCREAMING_SNAKE_CASE = bos_token_id def UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = None if self.use_input_mask: SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: SCREAMING_SNAKE_CASE = input_mask.numpy() SCREAMING_SNAKE_CASE = input_mask.shape SCREAMING_SNAKE_CASE = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowercase ): SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, tf.convert_to_tensor(_lowercase ) def UpperCamelCase ( self : List[str] ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def UpperCamelCase ( self : Dict , snake_case__ : str , snake_case__ : int , snake_case__ : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = TFBlipTextModel(config=_lowercase ) SCREAMING_SNAKE_CASE = model(_lowercase , attention_mask=_lowercase , training=_lowercase ) SCREAMING_SNAKE_CASE = model(_lowercase , training=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class UpperCamelCase ( a_ , unittest.TestCase ): __UpperCamelCase =(TFBlipTextModel,) if is_tf_available() else () __UpperCamelCase =False __UpperCamelCase =False __UpperCamelCase =False def UpperCamelCase ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE = BlipTextModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=_lowercase , hidden_size=3_7 ) def UpperCamelCase ( self : str ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCamelCase ( self : List[Any] ): """simple docstring""" pass def UpperCamelCase ( self : Optional[int] ): """simple docstring""" pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def UpperCamelCase ( self : Dict ): """simple docstring""" pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def UpperCamelCase ( self : int ): """simple docstring""" pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def UpperCamelCase ( self : Optional[int] ): """simple docstring""" pass @slow def UpperCamelCase ( self : str ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE = TFBlipTextModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def UpperCamelCase ( self : Union[str, Any] , snake_case__ : Tuple=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=_lowercase )
439
import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _UpperCAmelCase ( a__ , a__ , a__ , a__ , a__): '''simple docstring''' with open(a__) as metadata_file: a_ : Any = json.load(a__) a_ : Dict = LukeConfig(use_entity_aware_attention=a__ , **metadata["""model_config"""]) # Load in the weights from the checkpoint_path a_ : str = torch.load(a__ , map_location="""cpu""") # Load the entity vocab file a_ : List[str] = load_entity_vocab(a__) a_ : int = RobertaTokenizer.from_pretrained(metadata["""model_config"""]["""bert_model_name"""]) # Add special tokens to the token vocabulary for downstream tasks a_ : Optional[Any] = AddedToken("""<ent>""" , lstrip=a__ , rstrip=a__) a_ : int = AddedToken("""<ent2>""" , lstrip=a__ , rstrip=a__) tokenizer.add_special_tokens({"""additional_special_tokens""": [entity_token_a, entity_token_a]}) config.vocab_size += 2 print(f'''Saving tokenizer to {pytorch_dump_folder_path}''') tokenizer.save_pretrained(a__) with open(os.path.join(a__ , LukeTokenizer.vocab_files_names["""entity_vocab_file"""]) , """w""") as f: json.dump(a__ , a__) a_ : List[str] = LukeTokenizer.from_pretrained(a__) # Initialize the embeddings of the special tokens a_ : Optional[int] = state_dict["""embeddings.word_embeddings.weight"""] a_ : List[str] = word_emb[tokenizer.convert_tokens_to_ids(["""@"""])[0]].unsqueeze(0) a_ : List[Any] = word_emb[tokenizer.convert_tokens_to_ids(["""#"""])[0]].unsqueeze(0) a_ : Optional[Any] = torch.cat([word_emb, ent_emb, enta_emb]) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers): for matrix_name in ["query.weight", "query.bias"]: a_ : Any = f'''encoder.layer.{layer_index}.attention.self.''' a_ : List[str] = state_dict[prefix + matrix_name] a_ : List[Any] = state_dict[prefix + matrix_name] a_ : Dict = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks a_ : str = state_dict["""entity_embeddings.entity_embeddings.weight"""] a_ : int = entity_emb[entity_vocab["""[MASK]"""]] a_ : int = LukeModel(config=a__).eval() a_ , a_ : Optional[int] = model.load_state_dict(a__ , strict=a__) if not (len(a__) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(f'''Missing keys {", ".join(a__)}. Expected only missing embeddings.position_ids''') if not (all(key.startswith("""entity_predictions""") or key.startswith("""lm_head""") for key in unexpected_keys)): raise ValueError( """Unexpected keys""" f''' {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions") or key.startswith("lm_head"))])}''') # Check outputs a_ : List[Any] = LukeTokenizer.from_pretrained(a__ , task="""entity_classification""") a_ : str = ( """Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the""" """ new world number one avoid a humiliating second- round exit at Wimbledon .""" ) a_ : str = (3_9, 4_2) a_ : Tuple = tokenizer(a__ , entity_spans=[span] , add_prefix_space=a__ , return_tensors="""pt""") a_ : List[str] = model(**a__) # Verify word hidden states if model_size == "large": a_ : Optional[int] = torch.Size((1, 4_2, 1_0_2_4)) a_ : List[str] = torch.tensor( [[0.0133, 0.0865, 0.0095], [0.3093, -0.2576, -0.7418], [-0.1720, -0.2117, -0.2869]]) else: # base a_ : List[str] = torch.Size((1, 4_2, 7_6_8)) a_ : str = torch.tensor([[0.0037, 0.1368, -0.0091], [0.1099, 0.3329, -0.1095], [0.0765, 0.5335, 0.1179]]) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''') if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , a__ , atol=1e-4): raise ValueError # Verify entity hidden states if model_size == "large": a_ : Dict = torch.Size((1, 1, 1_0_2_4)) a_ : int = torch.tensor([[0.0466, -0.0106, -0.0179]]) else: # base a_ : Optional[Any] = torch.Size((1, 1, 7_6_8)) a_ : str = torch.tensor([[0.1457, 0.1044, 0.0174]]) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( f'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' f''' {expected_shape}''') if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , a__ , atol=1e-4): raise ValueError # Finally, save our PyTorch model and tokenizer print("""Saving PyTorch model to {}""".format(a__)) model.save_pretrained(a__) def _UpperCAmelCase ( a__): '''simple docstring''' a_ : List[str] = {} with open(a__ , """r""" , encoding="""utf-8""") as f: for index, line in enumerate(a__): a_ , a_ : List[Any] = line.rstrip().split("""\t""") a_ : Any = index return entity_vocab if __name__ == "__main__": __snake_case : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Path to a pytorch_model.bin file.""") parser.add_argument( """--metadata_path""", default=None, type=str, help="""Path to a metadata.json file, defining the configuration.""" ) parser.add_argument( """--entity_vocab_path""", default=None, type=str, help="""Path to an entity_vocab.tsv file, containing the entity vocabulary.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to where to dump the output PyTorch model.""" ) parser.add_argument( """--model_size""", default="""base""", type=str, choices=["""base""", """large"""], help="""Size of the model to be converted.""" ) __snake_case : List[str] = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
540
0
def _A ( lowerCamelCase ): return str(lowerCamelCase ) == str(lowerCamelCase )[::-1] def _A ( lowerCamelCase ): return int(lowerCamelCase ) + int(str(lowerCamelCase )[::-1] ) def _A ( lowerCamelCase = 1_0000 ): a__ : Any = [] for num in range(1 , lowerCamelCase ): a__ : int = 0 a__ : Any = num while iterations < 50: a__ : Tuple = sum_reverse(lowerCamelCase ) iterations += 1 if is_palindrome(lowerCamelCase ): break else: lychrel_nums.append(lowerCamelCase ) return len(lowerCamelCase ) if __name__ == "__main__": print(f'{solution() = }')
629
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __lowerCAmelCase ( _UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,unittest.TestCase ): _UpperCamelCase : Optional[int] = StableUnCLIPPipeline _UpperCamelCase : Optional[int] = TEXT_TO_IMAGE_PARAMS _UpperCamelCase : str = TEXT_TO_IMAGE_BATCH_PARAMS _UpperCamelCase : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS _UpperCamelCase : str = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false _UpperCamelCase : Any = False def _snake_case ( self ) -> List[str]: """simple docstring""" a__ : Any = 32 a__ : int = embedder_hidden_size # prior components torch.manual_seed(0 ) a__ : Any = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) a__ : Optional[Any] = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case , projection_dim=snake_case , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) ) torch.manual_seed(0 ) a__ : int = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=snake_case , num_layers=1 , ) torch.manual_seed(0 ) a__ : str = DDPMScheduler( variance_type="fixed_small_log" , prediction_type="sample" , num_train_timesteps=1_000 , clip_sample=snake_case , clip_sample_range=5.0 , beta_schedule="squaredcos_cap_v2" , ) # regular denoising components torch.manual_seed(0 ) a__ : Any = StableUnCLIPImageNormalizer(embedding_dim=snake_case ) a__ : int = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) a__ : Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) a__ : Union[str, Any] = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) ) torch.manual_seed(0 ) a__ : Any = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=snake_case , layers_per_block=1 , upcast_attention=snake_case , use_linear_projection=snake_case , ) torch.manual_seed(0 ) a__ : Tuple = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.00_085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=snake_case , steps_offset=1 , ) torch.manual_seed(0 ) a__ : Optional[int] = AutoencoderKL() a__ : Any = { # prior components "prior_tokenizer": prior_tokenizer, "prior_text_encoder": prior_text_encoder, "prior": prior, "prior_scheduler": prior_scheduler, # image noising components "image_normalizer": image_normalizer, "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder, "unet": unet, "scheduler": scheduler, "vae": vae, } return components def _snake_case ( self , snake_case , snake_case=0 ) -> Dict: """simple docstring""" if str(snake_case ).startswith("mps" ): a__ : Union[str, Any] = torch.manual_seed(snake_case ) else: a__ : List[str] = torch.Generator(device=snake_case ).manual_seed(snake_case ) a__ : Any = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "prior_num_inference_steps": 2, "output_type": "numpy", } return inputs def _snake_case ( self ) -> List[str]: """simple docstring""" a__ : Dict = torch_device == "cpu" self._test_attention_slicing_forward_pass(test_max_difference=snake_case ) def _snake_case ( self ) -> int: """simple docstring""" a__ : int = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=snake_case ) @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def _snake_case ( self ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ) -> Tuple: """simple docstring""" a__ : int = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy" ) a__ : int = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a__ : Dict = torch.Generator(device="cpu" ).manual_seed(0 ) a__ : Dict = pipe("anime turle" , generator=snake_case , output_type="np" ) a__ : List[str] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(snake_case , snake_case ) def _snake_case ( self ) -> Tuple: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() a__ : str = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) a__ : Union[str, Any] = pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a__ : Union[str, Any] = pipe( "anime turtle" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="np" , ) a__ : List[Any] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
629
1
import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class __A ( unittest.TestCase ): def _snake_case (self , __magic_name__ , __magic_name__ , __magic_name__ ): self.assertEqual(len(UpperCAmelCase_ ) , len(UpperCAmelCase_ ) ) for a, b in zip(UpperCAmelCase_ , UpperCAmelCase_ ): self.assertAlmostEqual(UpperCAmelCase_ , UpperCAmelCase_ , delta=UpperCAmelCase_ ) def _snake_case (self ): lowerCamelCase__ : Union[str, Any] = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(UpperCAmelCase_ ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step , 3 ) self.assertEqual(len(accumulator.gradients ) , 1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [-2.0, 5.0] , tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [0.0, 0.0] , tol=1E-2 ) def _snake_case (self ): lowerCamelCase__ : List[str] = None ops.enable_eager_execution_internal() lowerCamelCase__ : Optional[Any] = tf.config.list_physical_devices("""CPU""" ) if len(UpperCAmelCase_ ) == 1: tf.config.set_logical_device_configuration( physical_devices[0] , [tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) lowerCamelCase__ : List[str] = tf.config.list_logical_devices(device_type="""CPU""" ) lowerCamelCase__ : int = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): lowerCamelCase__ : List[str] = GradientAccumulator() lowerCamelCase__ : int = tf.Variable([4.0, 3.0] ) lowerCamelCase__ ,lowerCamelCase__ : List[Any] = create_optimizer(5E-5 , 10 , 5 ) lowerCamelCase__ : List[str] = tf.Variable([0.0, 0.0] , trainable=UpperCAmelCase_ ) def accumulate_on_replica(__magic_name__ ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients , [variable] ) ) ) @tf.function def accumulate(__magic_name__ , __magic_name__ ): with strategy.scope(): lowerCamelCase__ : List[str] = strategy.experimental_local_results(UpperCAmelCase_ ) local_variables[0].assign(UpperCAmelCase_ ) local_variables[1].assign(UpperCAmelCase_ ) strategy.run(UpperCAmelCase_ , args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(UpperCAmelCase_ ) def _check_local_values(__magic_name__ , __magic_name__ ): lowerCamelCase__ : Union[str, Any] = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value() , UpperCAmelCase_ , tol=1E-2 ) self.assertListAlmostEqual(values[1].value() , UpperCAmelCase_ , tol=1E-2 ) accumulate([1.0, 2.0] , [-1.0, 1.0] ) accumulate([3.0, -1.0] , [-1.0, -1.0] ) accumulate([-2.0, 2.0] , [3.0, -2.0] ) self.assertEqual(accumulator.step , 3 ) _check_local_values([2.0, 3.0] , [1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value() , [4.0, 3.0] , tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) _check_local_values([0.0, 0.0] , [0.0, 0.0] )
157
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case = { """configuration_blenderbot""": [ """BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlenderbotConfig""", """BlenderbotOnnxConfig""", ], """tokenization_blenderbot""": ["""BlenderbotTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""BlenderbotTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ """BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlenderbotForCausalLM""", """BlenderbotForConditionalGeneration""", """BlenderbotModel""", """BlenderbotPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ """TFBlenderbotForConditionalGeneration""", """TFBlenderbotModel""", """TFBlenderbotPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ """FlaxBlenderbotForConditionalGeneration""", """FlaxBlenderbotModel""", """FlaxBlenderbotPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
472
0
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class a__ ( a__ ): '''simple docstring''' lowercase__ : Tuple = DistilBertTokenizer lowercase__ : Any = DistilBertTokenizerFast lowercase__ : Dict = True @slow def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: lowerCAmelCase__ = DistilBertTokenizer.from_pretrained('''distilbert-base-uncased''' ) lowerCAmelCase__ = tokenizer.encode('''sequence builders''' , add_special_tokens=lowerCamelCase_ ) lowerCAmelCase__ = tokenizer.encode('''multi-sequence build''' , add_special_tokens=lowerCamelCase_ ) lowerCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ ) lowerCAmelCase__ = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ , lowerCamelCase_ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
98
'''simple docstring''' import doctest from collections import deque import numpy as np class a__ : '''simple docstring''' def __init__( self ) -> None: lowerCAmelCase__ = [2, 1, 2, -1] lowerCAmelCase__ = [1, 2, 3, 4] def __SCREAMING_SNAKE_CASE ( self ) -> list[float]: lowerCAmelCase__ = len(self.first_signal ) lowerCAmelCase__ = len(self.second_signal ) lowerCAmelCase__ = max(lowerCamelCase_ , lowerCamelCase_ ) # create a zero matrix of max_length x max_length lowerCAmelCase__ = [[0] * max_length for i in range(lowerCamelCase_ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(lowerCamelCase_ ): lowerCAmelCase__ = deque(self.second_signal ) rotated_signal.rotate(lowerCamelCase_ ) for j, item in enumerate(lowerCamelCase_ ): matrix[i][j] += item # multiply the matrix with the first signal lowerCAmelCase__ = np.matmul(np.transpose(lowerCamelCase_ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(lowerCamelCase_ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
98
1
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class _a ( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Optional[Any] = BertJapaneseTokenizer lowerCamelCase_ : str = False lowerCamelCase_ : List[Any] = True def __UpperCAmelCase( self ): super().setUp() __A : Union[str, Any] = [ "[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは", "世界", "##世界", "、", "##、", "。", "##。", ] __A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __UpperCAmelCase( self , __UpperCAmelCase ): __A : List[Any] = "こんにちは、世界。 \nこんばんは、世界。" __A : List[str] = "こんにちは 、 世界 。 こんばんは 、 世界 。" return input_text, output_text def __UpperCAmelCase( self , __UpperCAmelCase ): __A , __A : List[Any] = self.get_input_output_texts(__UpperCAmelCase ) __A : int = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __A : Any = tokenizer.decode(__UpperCAmelCase , clean_up_tokenization_spaces=__UpperCAmelCase ) return text, ids def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): __A : Optional[int] = self.tokenizer_class(self.vocab_file ) __A : Optional[int] = tokenizer.tokenize("こんにちは、世界。\nこんばんは、世界。" ) self.assertListEqual(__UpperCAmelCase , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __UpperCAmelCase( self ): __A : Dict = self.tokenizer_class(self.vocab_file , word_tokenizer_type="mecab" ) self.assertIsNotNone(__UpperCAmelCase ) __A : Optional[Any] = "こんにちは、世界。\nこんばんは、世界。" __A : Optional[int] = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A : List[str] = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(__UpperCAmelCase , "wb" ) as handle: pickle.dump(__UpperCAmelCase , __UpperCAmelCase ) with open(__UpperCAmelCase , "rb" ) as handle: __A : str = pickle.load(__UpperCAmelCase ) __A : Optional[Any] = tokenizer_new.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def __UpperCAmelCase( self ): __A : Any = MecabTokenizer(mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __UpperCAmelCase( self ): try: __A : Dict = MecabTokenizer(mecab_dic="unidic_lite" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __UpperCAmelCase( self ): try: __A : Dict = MecabTokenizer(mecab_dic="unidic" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __UpperCAmelCase( self ): __A : Dict = MecabTokenizer(do_lower_case=__UpperCAmelCase , mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iphone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __UpperCAmelCase( self ): try: __A : Dict = MecabTokenizer( do_lower_case=__UpperCAmelCase , normalize_text=__UpperCAmelCase , mecab_option="-d /usr/local/lib/mecab/dic/jumandic" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "\u3000", "。"] , ) def __UpperCAmelCase( self ): __A : int = MecabTokenizer(normalize_text=__UpperCAmelCase , mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", " ", "。"] , ) @require_sudachi def __UpperCAmelCase( self ): __A : List[Any] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="sudachi" ) self.assertIsNotNone(__UpperCAmelCase ) __A : Dict = "こんにちは、世界。\nこんばんは、世界。" __A : List[Any] = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A : int = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(__UpperCAmelCase , "wb" ) as handle: pickle.dump(__UpperCAmelCase , __UpperCAmelCase ) with open(__UpperCAmelCase , "rb" ) as handle: __A : Any = pickle.load(__UpperCAmelCase ) __A : Any = tokenizer_new.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @require_sudachi def __UpperCAmelCase( self ): __A : Optional[int] = SudachiTokenizer(sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def __UpperCAmelCase( self ): __A : Optional[Any] = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="A" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国", "人", "参政", "権"] ) @require_sudachi def __UpperCAmelCase( self ): __A : Optional[int] = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="B" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国人", "参政権"] ) @require_sudachi def __UpperCAmelCase( self ): __A : Any = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="C" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国人参政権"] ) @require_sudachi def __UpperCAmelCase( self ): __A : Tuple = SudachiTokenizer(do_lower_case=__UpperCAmelCase , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iphone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def __UpperCAmelCase( self ): __A : Dict = SudachiTokenizer(normalize_text=__UpperCAmelCase , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", "\u3000", "。", " ", " "] , ) @require_sudachi def __UpperCAmelCase( self ): __A : Optional[int] = SudachiTokenizer(trim_whitespace=__UpperCAmelCase , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) @require_jumanpp def __UpperCAmelCase( self ): __A : Any = self.tokenizer_class(self.vocab_file , word_tokenizer_type="jumanpp" ) self.assertIsNotNone(__UpperCAmelCase ) __A : Tuple = "こんにちは、世界。\nこんばんは、世界。" __A : int = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A : int = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(__UpperCAmelCase , "wb" ) as handle: pickle.dump(__UpperCAmelCase , __UpperCAmelCase ) with open(__UpperCAmelCase , "rb" ) as handle: __A : str = pickle.load(__UpperCAmelCase ) __A : Optional[int] = tokenizer_new.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @require_jumanpp def __UpperCAmelCase( self ): __A : str = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __UpperCAmelCase( self ): __A : int = JumanppTokenizer(do_lower_case=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iphone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __UpperCAmelCase( self ): __A : Dict = JumanppTokenizer(normalize_text=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["ア", "ッ", "フ", "゚", "ル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __UpperCAmelCase( self ): __A : Any = JumanppTokenizer(trim_whitespace=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "。"] , ) @require_jumanpp def __UpperCAmelCase( self ): __A : Tuple = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("ありがとうございますm(_ _)m見つけるのが大変です。" ) , ["ありがとう", "ございます", "m(_ _)m", "見つける", "の", "が", "大変です", "。"] , ) def __UpperCAmelCase( self ): __A : List[Any] = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは"] __A : Optional[int] = {} for i, token in enumerate(__UpperCAmelCase ): __A : List[str] = i __A : List[str] = WordpieceTokenizer(vocab=__UpperCAmelCase , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) , ["こんにちは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは" ) , ["こん", "##ばんは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは" ) , ["こん", "##ばんは", "[UNK]", "こんにちは"] ) def __UpperCAmelCase( self ): __A : Optional[Any] = BertJapaneseTokenizer.from_pretrained("nlp-waseda/roberta-base-japanese-with-auto-jumanpp" ) __A : List[str] = tokenizer.subword_tokenizer __A : int = subword_tokenizer.tokenize("国境 の 長い トンネル を 抜ける と 雪国 であった 。" ) self.assertListEqual(__UpperCAmelCase , ["▁国境", "▁の", "▁長い", "▁トンネル", "▁を", "▁抜ける", "▁と", "▁雪", "国", "▁であった", "▁。"] ) __A : Tuple = subword_tokenizer.tokenize("こんばんは こんばん にち は こんにちは" ) self.assertListEqual(__UpperCAmelCase , ["▁こん", "ばん", "は", "▁こん", "ばん", "▁に", "ち", "▁は", "▁こんにちは"] ) def __UpperCAmelCase( self ): __A : List[str] = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese" ) __A : Union[str, Any] = tokenizer.encode("ありがとう。" , add_special_tokens=__UpperCAmelCase ) __A : Dict = tokenizer.encode("どういたしまして。" , add_special_tokens=__UpperCAmelCase ) __A : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase ) __A : List[str] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase , __UpperCAmelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _a ( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Any = BertJapaneseTokenizer lowerCamelCase_ : int = False def __UpperCAmelCase( self ): super().setUp() __A : List[str] = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] __A : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __UpperCAmelCase( self , **__UpperCAmelCase ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="character" , **__UpperCAmelCase ) def __UpperCAmelCase( self , __UpperCAmelCase ): __A : Tuple = "こんにちは、世界。 \nこんばんは、世界。" __A : Optional[int] = "こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。" return input_text, output_text def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): pass # TODO add if relevant def __UpperCAmelCase( self ): __A : int = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="character" ) __A : Any = tokenizer.tokenize("こんにちは、世界。 \nこんばんは、世界。" ) self.assertListEqual( __UpperCAmelCase , ["こ", "ん", "に", "ち", "は", "、", "世", "界", "。", "こ", "ん", "ば", "ん", "は", "、", "世", "界", "。"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __UpperCAmelCase( self ): __A : Optional[int] = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] __A : Tuple = {} for i, token in enumerate(__UpperCAmelCase ): __A : Optional[Any] = i __A : Optional[int] = CharacterTokenizer(vocab=__UpperCAmelCase , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) , ["こ", "ん", "に", "ち", "は"] ) self.assertListEqual(tokenizer.tokenize("こんにちほ" ) , ["こ", "ん", "に", "ち", "[UNK]"] ) def __UpperCAmelCase( self ): __A : int = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese-char" ) __A : str = tokenizer.encode("ありがとう。" , add_special_tokens=__UpperCAmelCase ) __A : Union[str, Any] = tokenizer.encode("どういたしまして。" , add_special_tokens=__UpperCAmelCase ) __A : Optional[Any] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase ) __A : List[Any] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase , __UpperCAmelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _a ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase( self ): __A : Optional[int] = "cl-tohoku/bert-base-japanese" __A : Any = AutoTokenizer.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) class _a ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase( self ): __A : Tuple = "cl-tohoku/bert-base-japanese" with self.assertLogs("transformers" , level="WARNING" ) as cm: BertTokenizer.from_pretrained(__UpperCAmelCase ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) ) __A : Union[str, Any] = "bert-base-cased" with self.assertLogs("transformers" , level="WARNING" ) as cm: BertJapaneseTokenizer.from_pretrained(__UpperCAmelCase ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) )
520
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase = { 'configuration_ctrl': ['CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CTRLConfig'], 'tokenization_ctrl': ['CTRLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ 'CTRL_PRETRAINED_MODEL_ARCHIVE_LIST', 'CTRLForSequenceClassification', 'CTRLLMHeadModel', 'CTRLModel', 'CTRLPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ 'TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFCTRLForSequenceClassification', 'TFCTRLLMHeadModel', 'TFCTRLModel', 'TFCTRLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
520
1
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): A_ : Dict = "pt" elif is_tf_available(): A_ : Optional[int] = "tf" else: A_ : Tuple = "jax" class a_ ( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ : int = ByTaTokenizer lowerCamelCase__ : Optional[Any] = False def a__ (self ): '''simple docstring''' super().setUp() lowerCamelCase__ : Tuple = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def a__ (self ): '''simple docstring''' return ByTaTokenizer.from_pretrained('google/byt5-small' ) def a__ (self, **lowerCamelCase_ ): '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname, **_lowerCamelCase ) def a__ (self, lowerCamelCase_, lowerCamelCase_=False, lowerCamelCase_=2_0, lowerCamelCase_=5 ): '''simple docstring''' lowerCamelCase__ : int = [] for i in range(len(_lowerCamelCase ) ): try: lowerCamelCase__ : Any = tokenizer.decode([i], clean_up_tokenization_spaces=_lowerCamelCase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Tuple = list(filter(lambda lowerCamelCase_ : re.match(r'^[ a-zA-Z]+$', t[1] ), _lowerCamelCase ) ) lowerCamelCase__ : Union[str, Any] = list(filter(lambda lowerCamelCase_ : [t[0]] == tokenizer.encode(t[1], add_special_tokens=_lowerCamelCase ), _lowerCamelCase ) ) if max_length is not None and len(_lowerCamelCase ) > max_length: lowerCamelCase__ : str = toks[:max_length] if min_length is not None and len(_lowerCamelCase ) < min_length and len(_lowerCamelCase ) > 0: while len(_lowerCamelCase ) < min_length: lowerCamelCase__ : List[Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : List[Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : Optional[Any] = tokenizer.decode(_lowerCamelCase, clean_up_tokenization_spaces=_lowerCamelCase ) if " " not in output_txt and len(_lowerCamelCase ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]], clean_up_tokenization_spaces=_lowerCamelCase ) + ' ' + tokenizer.decode(toks_ids[1:], clean_up_tokenization_spaces=_lowerCamelCase ) ) if with_prefix_space: lowerCamelCase__ : List[str] = ' ' + output_txt lowerCamelCase__ : str = tokenizer.encode(_lowerCamelCase, add_special_tokens=_lowerCamelCase ) return output_txt, output_ids def a__ (self ): '''simple docstring''' lowerCamelCase__ : Tuple = self.ta_base_tokenizer lowerCamelCase__ : Union[str, Any] = tokenizer(['hi</s>', 'I went to the gym</s>', '</s>'] ) lowerCamelCase__ : Optional[int] = tokenizer(['hi', 'I went to the gym', ''] ) self.assertListEqual(batch_with_eos_added['input_ids'], batch_without_eos_added['input_ids'] ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : Optional[Any] = self.ta_base_tokenizer lowerCamelCase__ : Dict = 'Unicode €.' lowerCamelCase__ : List[str] = tokenizer(_lowerCamelCase ) lowerCamelCase__ : List[str] = [8_8, 1_1_3, 1_0_8, 1_0_2, 1_1_4, 1_0_3, 1_0_4, 3_5, 2_2_9, 1_3_3, 1_7_5, 4_9, 1] self.assertEqual(encoded['input_ids'], _lowerCamelCase ) # decoding lowerCamelCase__ : List[Any] = tokenizer.decode(_lowerCamelCase ) self.assertEqual(_lowerCamelCase, 'Unicode €.</s>' ) lowerCamelCase__ : Optional[Any] = tokenizer('e è é ê ë' ) lowerCamelCase__ : Dict = [1_0_4, 3_5, 1_9_8, 1_7_1, 3_5, 1_9_8, 1_7_2, 3_5, 1_9_8, 1_7_3, 3_5, 1_9_8, 1_7_4, 1] self.assertEqual(encoded['input_ids'], _lowerCamelCase ) # decoding lowerCamelCase__ : List[Any] = tokenizer.decode(_lowerCamelCase ) self.assertEqual(_lowerCamelCase, 'e è é ê ë</s>' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ), 'e è é ê ë</s>' ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : Any = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off lowerCamelCase__ : str = [6_8, 3_5, 1_1_1, 1_1_4, 1_1_3, 1_0_6, 3_5, 1_1_5, 1_0_0, 1_1_7, 1_0_0, 1_0_6, 1_1_7, 1_0_0, 1_1_5, 1_0_7, 3_5, 1_0_5, 1_1_4, 1_1_7, 3_5, 1_1_8, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_0_8, 1_2_5, 1_0_0, 1_1_9, 1_0_8, 1_1_4, 1_1_3, 4_9, 1, 0] # fmt: on lowerCamelCase__ : Dict = tokenizer(_lowerCamelCase, padding=_lowerCamelCase, return_tensors=_lowerCamelCase ) self.assertIsInstance(_lowerCamelCase, _lowerCamelCase ) if FRAMEWORK != "jax": lowerCamelCase__ : Optional[Any] = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : Optional[Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(_lowerCamelCase, _lowerCamelCase ) self.assertEqual((2, 3_7), batch.input_ids.shape ) self.assertEqual((2, 3_7), batch.attention_mask.shape ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : Dict = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] lowerCamelCase__ : List[Any] = tokenizer(_lowerCamelCase, padding=_lowerCamelCase, return_tensors=_lowerCamelCase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids', _lowerCamelCase ) self.assertIn('attention_mask', _lowerCamelCase ) self.assertNotIn('decoder_input_ids', _lowerCamelCase ) self.assertNotIn('decoder_attention_mask', _lowerCamelCase ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : Union[str, Any] = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = [ 'Summary of the text.', 'Another summary.', ] lowerCamelCase__ : List[str] = tokenizer( text_target=_lowerCamelCase, max_length=3_2, padding='max_length', truncation=_lowerCamelCase, return_tensors=_lowerCamelCase ) self.assertEqual(3_2, targets['input_ids'].shape[1] ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : List[str] = ['A long paragraph for summarization. </s>'] lowerCamelCase__ : int = ['Summary of the text. </s>'] # fmt: off lowerCamelCase__ : Dict = [6_8, 3_5, 1_1_1, 1_1_4, 1_1_3, 1_0_6, 3_5, 1_1_5, 1_0_0, 1_1_7, 1_0_0, 1_0_6, 1_1_7, 1_0_0, 1_1_5, 1_0_7, 3_5, 1_0_5, 1_1_4, 1_1_7, 3_5, 1_1_8, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_0_8, 1_2_5, 1_0_0, 1_1_9, 1_0_8, 1_1_4, 1_1_3, 4_9, 3_5, 1] lowerCamelCase__ : Optional[int] = [8_6, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_2_4, 3_5, 1_1_4, 1_0_5, 3_5, 1_1_9, 1_0_7, 1_0_4, 3_5, 1_1_9, 1_0_4, 1_2_3, 1_1_9, 4_9, 3_5, 1] # fmt: on lowerCamelCase__ : Optional[int] = tokenizer(_lowerCamelCase, text_target=_lowerCamelCase ) self.assertEqual(_lowerCamelCase, batch['input_ids'][0] ) self.assertEqual(_lowerCamelCase, batch['labels'][0] ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length, 4_2 ) # Now let's start the test lowerCamelCase__ : Tuple = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : List[str] = ' He is very happy, UNwant\u00E9d,running' lowerCamelCase__ : Optional[int] = tokenizer.encode(_lowerCamelCase, add_special_tokens=_lowerCamelCase ) tokenizer.save_pretrained(_lowerCamelCase ) lowerCamelCase__ : Tuple = tokenizer.__class__.from_pretrained(_lowerCamelCase ) lowerCamelCase__ : int = after_tokenizer.encode(_lowerCamelCase, add_special_tokens=_lowerCamelCase ) self.assertListEqual(_lowerCamelCase, _lowerCamelCase ) shutil.rmtree(_lowerCamelCase ) lowerCamelCase__ : List[str] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Union[str, Any] = tempfile.mkdtemp() lowerCamelCase__ : Dict = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) lowerCamelCase__ : Dict = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) lowerCamelCase__ : Optional[int] = tokenizer.encode(_lowerCamelCase, add_special_tokens=_lowerCamelCase ) tokenizer.save_pretrained(_lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(_lowerCamelCase ) lowerCamelCase__ : int = after_tokenizer.encode(_lowerCamelCase, add_special_tokens=_lowerCamelCase ) self.assertListEqual(_lowerCamelCase, _lowerCamelCase ) self.assertIn('new_additional_special_token', after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length, 4_2 ) lowerCamelCase__ : Optional[Any] = tokenizer.__class__.from_pretrained(_lowerCamelCase, model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length, 4_3 ) shutil.rmtree(_lowerCamelCase ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_lowerCamelCase ) with open(os.path.join(_lowerCamelCase, 'special_tokens_map.json' ), encoding='utf-8' ) as json_file: lowerCamelCase__ : Union[str, Any] = json.load(_lowerCamelCase ) with open(os.path.join(_lowerCamelCase, 'tokenizer_config.json' ), encoding='utf-8' ) as json_file: lowerCamelCase__ : Optional[int] = json.load(_lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = [f'''<extra_id_{i}>''' for i in range(1_2_5 )] lowerCamelCase__ : List[Any] = added_tokens_extra_ids + [ 'an_additional_special_token' ] lowerCamelCase__ : Dict = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(_lowerCamelCase, 'special_tokens_map.json' ), 'w', encoding='utf-8' ) as outfile: json.dump(_lowerCamelCase, _lowerCamelCase ) with open(os.path.join(_lowerCamelCase, 'tokenizer_config.json' ), 'w', encoding='utf-8' ) as outfile: json.dump(_lowerCamelCase, _lowerCamelCase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : int = tokenizer_class.from_pretrained( _lowerCamelCase, ) self.assertIn( 'an_additional_special_token', tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['an_additional_special_token'], tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ), ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : List[Any] = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token', lstrip=_lowerCamelCase )] lowerCamelCase__ : Optional[int] = tokenizer_class.from_pretrained( _lowerCamelCase, additional_special_tokens=_lowerCamelCase, ) self.assertIn('a_new_additional_special_token', tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'], tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ), ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : str = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_lowerCamelCase ) lowerCamelCase__ : Any = tokenizer_class.from_pretrained(_lowerCamelCase ) self.assertTrue(tokenizer.decode([2_5_5] ) == '' ) def a__ (self ): '''simple docstring''' pass def a__ (self ): '''simple docstring''' pass def a__ (self ): '''simple docstring''' pass def a__ (self ): '''simple docstring''' pass def a__ (self ): '''simple docstring''' lowerCamelCase__ : Dict = self.get_tokenizers(fast=_lowerCamelCase, do_lower_case=_lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Optional[Any] = ['t', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 'x', 't', '</s>'] lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_string(_lowerCamelCase ) self.assertIsInstance(_lowerCamelCase, _lowerCamelCase ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : List[str] = [ 'bos_token', 'eos_token', 'unk_token', 'sep_token', 'pad_token', 'cls_token', 'mask_token', ] lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : List[Any] = tokenizer.convert_ids_to_tokens( _lowerCamelCase, skip_special_tokens=_lowerCamelCase ) for attr in attributes_list: setattr(_lowerCamelCase, attr + '_id', _lowerCamelCase ) self.assertEqual(getattr(_lowerCamelCase, _lowerCamelCase ), _lowerCamelCase ) self.assertEqual(getattr(_lowerCamelCase, attr + '_id' ), _lowerCamelCase ) setattr(_lowerCamelCase, attr + '_id', _lowerCamelCase ) self.assertEqual(getattr(_lowerCamelCase, _lowerCamelCase ), _lowerCamelCase ) self.assertEqual(getattr(_lowerCamelCase, attr + '_id' ), _lowerCamelCase ) setattr(_lowerCamelCase, 'additional_special_tokens_ids', [] ) self.assertListEqual(getattr(_lowerCamelCase, 'additional_special_tokens' ), [] ) self.assertListEqual(getattr(_lowerCamelCase, 'additional_special_tokens_ids' ), [] ) setattr(_lowerCamelCase, 'additional_special_tokens_ids', [token_id_to_test_setters] ) self.assertListEqual(getattr(_lowerCamelCase, 'additional_special_tokens' ), [token_to_test_setters] ) self.assertListEqual(getattr(_lowerCamelCase, 'additional_special_tokens_ids' ), [token_id_to_test_setters] )
712
"""simple docstring""" A_ : List[str] = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", "dataclasses": "dataclasses", "datasets": "datasets!=2.5.0", "decord": "decord==0.6.0", "deepspeed": "deepspeed>=0.9.3", "diffusers": "diffusers", "dill": "dill<0.3.5", "evaluate": "evaluate>=0.2.0", "fairscale": "fairscale>0.3", "faiss-cpu": "faiss-cpu", "fastapi": "fastapi", "filelock": "filelock", "flax": "flax>=0.4.1,<=0.7.0", "ftfy": "ftfy", "fugashi": "fugashi>=1.0", "GitPython": "GitPython<3.1.19", "hf-doc-builder": "hf-doc-builder>=0.3.0", "huggingface-hub": "huggingface-hub>=0.14.1,<1.0", "importlib_metadata": "importlib_metadata", "ipadic": "ipadic>=1.0.0,<2.0", "isort": "isort>=5.5.4", "jax": "jax>=0.2.8,!=0.3.2,<=0.4.13", "jaxlib": "jaxlib>=0.1.65,<=0.4.13", "jieba": "jieba", "kenlm": "kenlm", "keras-nlp": "keras-nlp>=0.3.1", "librosa": "librosa", "nltk": "nltk", "natten": "natten>=0.14.6", "numpy": "numpy>=1.17", "onnxconverter-common": "onnxconverter-common", "onnxruntime-tools": "onnxruntime-tools>=1.4.2", "onnxruntime": "onnxruntime>=1.4.0", "opencv-python": "opencv-python", "optuna": "optuna", "optax": "optax>=0.0.8,<=0.1.4", "packaging": "packaging>=20.0", "parameterized": "parameterized", "phonemizer": "phonemizer", "protobuf": "protobuf", "psutil": "psutil", "pyyaml": "pyyaml>=5.1", "pydantic": "pydantic<2", "pytest": "pytest>=7.2.0", "pytest-timeout": "pytest-timeout", "pytest-xdist": "pytest-xdist", "python": "python>=3.8.0", "ray[tune]": "ray[tune]", "regex": "regex!=2019.12.17", "requests": "requests", "rhoknp": "rhoknp>=1.1.0,<1.3.1", "rjieba": "rjieba", "rouge-score": "rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1", "ruff": "ruff>=0.0.241,<=0.0.259", "sacrebleu": "sacrebleu>=1.4.12,<2.0.0", "sacremoses": "sacremoses", "safetensors": "safetensors>=0.3.1", "sagemaker": "sagemaker>=2.31.0", "scikit-learn": "scikit-learn", "sentencepiece": "sentencepiece>=0.1.91,!=0.1.92", "sigopt": "sigopt", "starlette": "starlette", "sudachipy": "sudachipy>=0.6.6", "sudachidict_core": "sudachidict_core>=20220729", "tensorflow-cpu": "tensorflow-cpu>=2.6,<2.14", "tensorflow": "tensorflow>=2.6,<2.14", "tensorflow-text": "tensorflow-text<2.14", "tf2onnx": "tf2onnx", "timeout-decorator": "timeout-decorator", "timm": "timm", "tokenizers": "tokenizers>=0.11.1,!=0.11.3,<0.14", "torch": "torch>=1.9,!=1.12.0", "torchaudio": "torchaudio", "torchvision": "torchvision", "pyctcdecode": "pyctcdecode>=0.4.0", "tqdm": "tqdm>=4.27", "unidic": "unidic>=1.0.2", "unidic_lite": "unidic_lite>=1.0.7", "urllib3": "urllib3<2.0.0", "uvicorn": "uvicorn", }
696
0
from typing import Union import fire import torch from tqdm import tqdm def SCREAMING_SNAKE_CASE ( snake_case__ , snake_case__ = "cpu" , snake_case__ = None ) -> None: __UpperCAmelCase =torch.load(snake_case__ , map_location=snake_case__ ) for k, v in tqdm(state_dict.items() ): if not isinstance(snake_case__ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) __UpperCAmelCase =v.half() if save_path is None: # overwrite src_path __UpperCAmelCase =src_path torch.save(snake_case__ , snake_case__ ) if __name__ == "__main__": fire.Fire(convert)
132
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def SCREAMING_SNAKE_CASE ( snake_case__=32 , snake_case__=10 , snake_case__=100 , snake_case__=1026 , snake_case__=True , snake_case__="data/tokenized_stories_train_wikitext103.jbl" , snake_case__="igf_context_pairs.jbl" , ) -> Any: set_seed(3 ) # generate train_data and objective_set __UpperCAmelCase , __UpperCAmelCase =generate_datasets( snake_case__ , snake_case__ , number=snake_case__ , min_len=1026 , trim=snake_case__ ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? __UpperCAmelCase =torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) # load pretrained model __UpperCAmelCase =load_gpta('''gpt2''' ).to(snake_case__ ) print('''computing perplexity on objective set''' ) __UpperCAmelCase =compute_perplexity(snake_case__ , snake_case__ , snake_case__ ).item() print('''perplexity on objective set:''' , snake_case__ ) # collect igf pairs and save to file demo.jbl collect_objective_set(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( snake_case__ , snake_case__=15 , snake_case__=128 , snake_case__=100 , snake_case__="igf_model.pt" , ) -> str: set_seed(42 ) # Load pre-trained model __UpperCAmelCase =GPTaLMHeadModel.from_pretrained('''gpt2''' ) # Initialize secondary learner to use embedding weights of model __UpperCAmelCase =SecondaryLearner(snake_case__ ) # Train secondary learner __UpperCAmelCase =train_secondary_learner( snake_case__ , snake_case__ , max_epochs=snake_case__ , batch_size=snake_case__ , eval_freq=100 , igf_model_path=snake_case__ , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def SCREAMING_SNAKE_CASE ( snake_case__ , snake_case__ , snake_case__ , snake_case__=32 , snake_case__=1000 , snake_case__=16 , snake_case__=1.0 , snake_case__=recopy_gpta , snake_case__=None , snake_case__=10 , snake_case__="gpt2_finetuned.pt" , ) -> List[str]: __UpperCAmelCase =torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) __UpperCAmelCase =RandomSampler(snake_case__ ) __UpperCAmelCase =DataLoader(snake_case__ , sampler=snake_case__ ) __UpperCAmelCase =max_steps // (len(snake_case__ )) + 1 __UpperCAmelCase =0 __UpperCAmelCase =torch.zeros((1, context_len) , dtype=torch.long , device=snake_case__ ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =recopy_model(snake_case__ , snake_case__ , snake_case__ ) model.train() if secondary_learner is not None: secondary_learner.to(snake_case__ ) secondary_learner.eval() __UpperCAmelCase =[] __UpperCAmelCase =0 __UpperCAmelCase =[] __UpperCAmelCase =[] # Compute the performance of the transformer model at the beginning __UpperCAmelCase =compute_perplexity(snake_case__ , snake_case__ , snake_case__ ) test_perps.append(snake_case__ ) print('''Test perplexity, step''' , snake_case__ , ''':''' , snake_case__ ) for epoch in range(int(snake_case__ ) ): for step, example in enumerate(snake_case__ ): torch.cuda.empty_cache() __UpperCAmelCase =random.randint(0 , example.size(2 ) - context_len - 1 ) __UpperCAmelCase =example[0, 0, start : start + context_len] lm_optimizer.zero_grad() __UpperCAmelCase =model(snake_case__ , labels=snake_case__ ) __UpperCAmelCase =True if secondary_learner is not None: __UpperCAmelCase =secondary_learner.forward( torch.tensor(snake_case__ , dtype=torch.long , device=snake_case__ ).unsqueeze(0 ) )[0].item() observed_qs.append(float(snake_case__ ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: __UpperCAmelCase =-1 if predicted_q < threshold: __UpperCAmelCase =False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) __UpperCAmelCase =outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() __UpperCAmelCase =0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: __UpperCAmelCase =compute_perplexity(snake_case__ , snake_case__ , snake_case__ ) test_perps.append(snake_case__ ) print('''Test perplexity, step''' , snake_case__ , ''':''' , snake_case__ ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , snake_case__ ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def SCREAMING_SNAKE_CASE ( ) -> List[Any]: __UpperCAmelCase =argparse.ArgumentParser(description='''Fine-tune a transformer model with IGF on a language modeling task''' ) # Required parameters parser.add_argument( '''--data_dir''' , default=snake_case__ , type=snake_case__ , required=snake_case__ , help='''The input data dir. Should contain data files for WikiText.''' , ) parser.add_argument( '''--model_name_or_path''' , default=snake_case__ , type=snake_case__ , required=snake_case__ , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--data_file''' , type=snake_case__ , default=snake_case__ , help=( '''A jbl file containing tokenized data which can be split as objective dataset, ''' '''train_dataset and test_dataset.''' ) , ) parser.add_argument( '''--igf_data_file''' , type=snake_case__ , default=snake_case__ , help='''A jbl file containing the context and information gain pairs to train secondary learner.''' , ) parser.add_argument( '''--output_dir''' , default=snake_case__ , type=snake_case__ , required=snake_case__ , help='''The output directory where the final fine-tuned model is stored.''' , ) parser.add_argument( '''--tokenizer_name''' , default=snake_case__ , type=snake_case__ , help='''Pretrained tokenizer name or path if not the same as model_name''' , ) parser.add_argument('''--seed''' , type=snake_case__ , default=snake_case__ , help='''A seed for reproducible training.''' ) parser.add_argument( '''--context_len''' , default=32 , type=snake_case__ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--size_objective_set''' , default=100 , type=snake_case__ , help='''number of articles that are long enough to be used as our objective set''' , ) parser.add_argument( '''--eval_freq''' , default=100 , type=snake_case__ , help='''secondary model evaluation is triggered at eval_freq''' ) parser.add_argument('''--max_steps''' , default=1000 , type=snake_case__ , help='''To calculate training epochs''' ) parser.add_argument( '''--secondary_learner_batch_size''' , default=128 , type=snake_case__ , help='''batch size of training data for secondary learner''' , ) parser.add_argument( '''--batch_size''' , default=16 , type=snake_case__ , help='''batch size of training data of language model(gpt2) ''' ) parser.add_argument( '''--eval_interval''' , default=10 , type=snake_case__ , help=( '''decay the selectivity of our secondary learner filter from''' '''1 standard deviation above average to 1 below average after 10 batches''' ) , ) parser.add_argument( '''--number''' , default=100 , type=snake_case__ , help='''The number of examples split to be used as objective_set/test_data''' ) parser.add_argument( '''--min_len''' , default=1026 , type=snake_case__ , help='''The minimum length of the article to be used as objective set''' ) parser.add_argument( '''--secondary_learner_max_epochs''' , default=15 , type=snake_case__ , help='''number of epochs to train secondary learner''' ) parser.add_argument('''--trim''' , default=snake_case__ , type=snake_case__ , help='''truncate the example if it exceeds context length''' ) parser.add_argument( '''--threshold''' , default=1.0 , type=snake_case__ , help=( '''The threshold value used by secondary learner to filter the train_data and allow only''' ''' informative data as input to the model''' ) , ) parser.add_argument('''--finetuned_model_name''' , default='''gpt2_finetuned.pt''' , type=snake_case__ , help='''finetuned_model_name''' ) parser.add_argument( '''--recopy_model''' , default=snake_case__ , type=snake_case__ , help='''Reset the model to the original pretrained GPT-2 weights after each iteration''' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1026 , trim=snake_case__ , data_file='''data/tokenized_stories_train_wikitext103.jbl''' , igf_data_file='''igf_context_pairs.jbl''' , ) # Load train data for secondary learner __UpperCAmelCase =joblib.load('''data/IGF_values.jbl''' ) # Train secondary learner __UpperCAmelCase =training_secondary_learner( snake_case__ , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='''igf_model.pt''' , ) # load pretrained gpt2 model __UpperCAmelCase =GPTaLMHeadModel.from_pretrained('''gpt2''' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model __UpperCAmelCase , __UpperCAmelCase =generate_datasets( context_len=32 , file='''data/tokenized_stories_train_wikitext103.jbl''' , number=100 , min_len=1026 , trim=snake_case__ ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( snake_case__ , snake_case__ , snake_case__ , context_len=32 , max_steps=1000 , batch_size=16 , threshold=1.0 , recopy_model=snake_case__ , secondary_learner=snake_case__ , eval_interval=10 , finetuned_model_name='''gpt2_finetuned.pt''' , ) if __name__ == "__main__": main()
132
1
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() snake_case : Optional[int] = logging.get_logger(__name__) snake_case : int = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def lowercase__ ( __UpperCamelCase : List[str] , __UpperCamelCase : Any , __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : str ): '''simple docstring''' for attribute in key.split(""".""" ): __lowercase = getattr(__UpperCamelCase , __UpperCamelCase ) if weight_type is not None: __lowercase = getattr(__UpperCamelCase , __UpperCamelCase ).shape else: __lowercase = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": __lowercase = value elif weight_type == "weight_g": __lowercase = value elif weight_type == "weight_v": __lowercase = value elif weight_type == "bias": __lowercase = value else: __lowercase = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowercase__ ( __UpperCamelCase : Dict , __UpperCamelCase : int , __UpperCamelCase : Optional[int] ): '''simple docstring''' __lowercase = [] __lowercase = fairseq_model.state_dict() __lowercase = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __lowercase = False if "conv_layers" in name: load_conv_layer( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , hf_model.config.feat_extract_norm == """group""" , ) __lowercase = True else: for key, mapped_key in MAPPING.items(): __lowercase = """hubert.""" + mapped_key if (is_finetuned and mapped_key != """lm_head""") else mapped_key if key in name or (key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0] and not is_finetuned): __lowercase = True if "*" in mapped_key: __lowercase = name.split(__UpperCamelCase )[0].split(""".""" )[-2] __lowercase = mapped_key.replace("""*""" , __UpperCamelCase ) if "weight_g" in name: __lowercase = """weight_g""" elif "weight_v" in name: __lowercase = """weight_v""" elif "weight" in name: __lowercase = """weight""" elif "bias" in name: __lowercase = """bias""" else: __lowercase = None set_recursively(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) continue if not is_used: unused_weights.append(__UpperCamelCase ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase__ ( __UpperCamelCase : Optional[int] , __UpperCamelCase : int , __UpperCamelCase : List[str] , __UpperCamelCase : Dict , __UpperCamelCase : Union[str, Any] ): '''simple docstring''' __lowercase = full_name.split("""conv_layers.""" )[-1] __lowercase = name.split(""".""" ) __lowercase = int(items[0] ) __lowercase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) __lowercase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) __lowercase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) __lowercase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) __lowercase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__UpperCamelCase ) @torch.no_grad() def lowercase__ ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : Any , __UpperCamelCase : Optional[Any]=None , __UpperCamelCase : List[str]=None , __UpperCamelCase : Optional[Any]=True ): '''simple docstring''' if config_path is not None: __lowercase = HubertConfig.from_pretrained(__UpperCamelCase ) else: __lowercase = HubertConfig() if is_finetuned: if dict_path: __lowercase = Dictionary.load(__UpperCamelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowercase = target_dict.pad_index __lowercase = target_dict.bos_index __lowercase = target_dict.eos_index __lowercase = len(target_dict.symbols ) __lowercase = os.path.join(__UpperCamelCase , """vocab.json""" ) if not os.path.isdir(__UpperCamelCase ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(__UpperCamelCase ) ) return os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(target_dict.indices , __UpperCamelCase ) __lowercase = WavaVecaCTCTokenizer( __UpperCamelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=__UpperCamelCase , ) __lowercase = True if config.feat_extract_norm == """layer""" else False __lowercase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=__UpperCamelCase , return_attention_mask=__UpperCamelCase , ) __lowercase = WavaVecaProcessor(feature_extractor=__UpperCamelCase , tokenizer=__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) __lowercase = HubertForCTC(__UpperCamelCase ) else: __lowercase = HubertModel(__UpperCamelCase ) if is_finetuned: __lowercase , __lowercase , __lowercase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) else: __lowercase , __lowercase , __lowercase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __lowercase = model[0].eval() recursively_load_weights(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) hf_wavavec.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": snake_case : str = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) snake_case : Optional[int] = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
705
'''simple docstring''' import re from filelock import FileLock try: import nltk snake_case : Optional[Any] = True except (ImportError, ModuleNotFoundError): snake_case : Dict = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def lowercase__ ( __UpperCamelCase : str ): '''simple docstring''' re.sub("""<n>""" , """""" , __UpperCamelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__UpperCamelCase ) )
339
0
from abc import ABC, abstractmethod from argparse import ArgumentParser class lowerCAmelCase ( UpperCAmelCase_ ): @staticmethod @abstractmethod def A_ ( UpperCAmelCase : ArgumentParser ) -> str: raise NotImplementedError() @abstractmethod def A_ ( self : Union[str, Any] ) -> Dict: raise NotImplementedError()
295
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase : """simple docstring""" def __init__( self : Dict , _snake_case : Union[str, Any] , _snake_case : Optional[Any]=12 , _snake_case : Any=7 , _snake_case : List[str]=True , _snake_case : int=True , _snake_case : int=True , _snake_case : Tuple=99 , _snake_case : List[Any]=32 , _snake_case : Optional[int]=32 , _snake_case : List[str]=2 , _snake_case : List[str]=4 , _snake_case : List[Any]=37 , _snake_case : Union[str, Any]=0.1 , _snake_case : Tuple=0.1 , _snake_case : Dict=5_12 , _snake_case : Union[str, Any]=0.02 , _snake_case : Any=0 , _snake_case : Optional[Any]=None , ): """simple docstring""" A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = projection_dim A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = dropout A__ = attention_dropout A__ = max_position_embeddings A__ = initializer_range A__ = scope A__ = bos_token_id def _a ( self : Optional[Any] ): """simple docstring""" A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: A__ = input_mask.numpy() A__ , A__ = input_mask.shape A__ = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_snake_case ): A__ = 1 A__ = 0 A__ = self.get_config() return config, input_ids, tf.convert_to_tensor(_snake_case ) def _a ( self : Tuple ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _a ( self : int , _snake_case : Union[str, Any] , _snake_case : Any , _snake_case : List[str] ): """simple docstring""" A__ = TFBlipTextModel(config=_snake_case ) A__ = model(_snake_case , attention_mask=_snake_case , training=_snake_case ) A__ = model(_snake_case , training=_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a ( self : str ): """simple docstring""" A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" A__ : Tuple = (TFBlipTextModel,) if is_tf_available() else () A__ : Optional[int] = False A__ : Union[str, Any] = False A__ : Union[str, Any] = False def _a ( self : Any ): """simple docstring""" A__ = BlipTextModelTester(self ) A__ = ConfigTester(self , config_class=_snake_case , hidden_size=37 ) def _a ( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Union[str, Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def _a ( self : Tuple ): """simple docstring""" pass def _a ( self : int ): """simple docstring""" pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _a ( self : Any ): """simple docstring""" pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _a ( self : str ): """simple docstring""" pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _a ( self : Optional[Any] ): """simple docstring""" pass @slow def _a ( self : Union[str, Any] ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = TFBlipTextModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def _a ( self : int , _snake_case : int=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=_snake_case )
9
0
def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ = 100 ): __lowerCamelCase : Dict = 0 __lowerCamelCase : Union[str, Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
230
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'vocab.txt', 'emoji_file': 'emoji.json'} lowercase_ = { 'vocab_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt', }, 'emoji_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json', }, } lowercase_ = { 'abeja/gpt-neox-japanese-2.7b': 2_0_4_8, } def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): with open(SCREAMING_SNAKE_CASE__ , 'r' , encoding='utf-8' ) as f: __lowerCamelCase : Any = json.loads(f.read() ) __lowerCamelCase : Optional[int] = collections.OrderedDict() __lowerCamelCase : List[Any] = collections.OrderedDict() __lowerCamelCase : int = collections.OrderedDict() with open(SCREAMING_SNAKE_CASE__ , 'r' , encoding='utf-8' ) as f: __lowerCamelCase : Dict = f.readlines() __lowerCamelCase : Optional[int] = [[t.rstrip('\n' )] if (t == ',' or ',' not in t) else t.rstrip('\n' ).split(',' ) for t in token] for idx, b in enumerate(SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Union[str, Any] = b __lowerCamelCase : List[str] = idx for wd in b: __lowerCamelCase : Union[str, Any] = idx return vocab, raw_vocab, ids_to_tokens, emoji class A_ ( __UpperCamelCase ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = ["""input_ids""", """attention_mask"""] def __init__( self: str , a: Tuple , a: List[str] , a: Optional[Any]="<|endoftext|>" , a: Dict="<|endoftext|>" , a: Tuple="<|startoftext|>" , a: Any="<|endoftext|>" , a: Dict=False , **a: Union[str, Any] , ): super().__init__( unk_token=a , pad_token=a , bos_token=a , eos_token=a , do_clean_text=a , **a , ) if not os.path.isfile(a ): raise ValueError( F'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' ' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) if not os.path.isfile(a ): raise ValueError( F'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' ' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) __lowerCamelCase : Optional[int] = do_clean_text __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : List[str] = load_vocab_and_emoji(a , a ) __lowerCamelCase : List[str] = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def _snake_case ( self: Dict ): # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def _snake_case ( self: Any ): return dict(self.raw_vocab , **self.added_tokens_encoder ) def _snake_case ( self: str , a: List[str] ): return self.subword_tokenizer.tokenize(a , clean=self.do_clean_text ) def _snake_case ( self: Tuple , a: Optional[Any] ): return self.vocab.get(a , self.vocab.get(self.unk_token ) ) def _snake_case ( self: Union[str, Any] , a: Tuple ): return self.subword_tokenizer.convert_id_to_token(a ) def _snake_case ( self: Optional[int] , a: str ): __lowerCamelCase : int = ''.join(a ).strip() return out_string def _snake_case ( self: Tuple , a: "Conversation" ): __lowerCamelCase : Optional[int] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(a , add_special_tokens=a ) + [self.eos_token_id] ) if len(a ) > self.model_max_length: __lowerCamelCase : int = input_ids[-self.model_max_length :] return input_ids def _snake_case ( self: Optional[Any] , a: str , a: Optional[str] = None ): __lowerCamelCase : int = 0 if os.path.isdir(a ): __lowerCamelCase : Tuple = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) __lowerCamelCase : Tuple = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['emoji_file'] ) else: __lowerCamelCase : List[str] = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) __lowerCamelCase : Optional[int] = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(a , 'w' , encoding='utf-8' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ' Please check that the vocabulary is not corrupted!' ) __lowerCamelCase : int = token_index writer.write(','.join(a ) + '\n' ) index += 1 with open(a , 'w' , encoding='utf-8' ) as writer: json.dump(self.emoji , a ) return vocab_file, emoji_file class A_ ( __UpperCamelCase ): '''simple docstring''' def __init__( self: List[str] , a: str , a: Union[str, Any] , a: int ): __lowerCamelCase : int = vocab # same as swe __lowerCamelCase : int = ids_to_tokens # same as bpe __lowerCamelCase : Tuple = emoji __lowerCamelCase : int = np.max([len(a ) for w in self.vocab.keys()] ) __lowerCamelCase : Optional[int] = re.compile(R'(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)' ) __lowerCamelCase : Optional[Any] = re.compile(R'[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*' ) __lowerCamelCase : Dict = re.compile(R'[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}' ) __lowerCamelCase : int = re.compile( R'([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) __lowerCamelCase : List[str] = re.compile( R'(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) __lowerCamelCase : Optional[int] = re.compile( R'((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*' ) __lowerCamelCase : str = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' __lowerCamelCase : str = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' __lowerCamelCase : Optional[int] = str.maketrans({k: '<BLOCK>' for k in keisen + blocks} ) def __len__( self: Optional[int] ): return len(self.ids_to_tokens ) def _snake_case ( self: Any , a: Tuple ): __lowerCamelCase : Optional[Any] = self.content_repattera.sub('<URL>' , a ) __lowerCamelCase : Optional[int] = self.content_repattera.sub('<EMAIL>' , a ) __lowerCamelCase : int = self.content_repattera.sub('<TEL>' , a ) __lowerCamelCase : Optional[Any] = self.content_repattera.sub('<DATE>' , a ) __lowerCamelCase : Optional[Any] = self.content_repattera.sub('<DATE>' , a ) __lowerCamelCase : str = self.content_repattera.sub('<PRICE>' , a ) __lowerCamelCase : str = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __lowerCamelCase : int = content.replace('<BLOCK><BLOCK>' , '<BLOCK>' ) return content def _snake_case ( self: Any , a: Union[str, Any] , a: Tuple=False ): __lowerCamelCase : List[Any] = text.replace(' ' , '<SP>' ) __lowerCamelCase : Union[str, Any] = text.replace(' ' , '<SP>' ) __lowerCamelCase : Any = text.replace('\r\n' , '<BR>' ) __lowerCamelCase : List[str] = text.replace('\n' , '<BR>' ) __lowerCamelCase : Optional[Any] = text.replace('\r' , '<BR>' ) __lowerCamelCase : str = text.replace('\t' , '<TAB>' ) __lowerCamelCase : List[str] = text.replace('—' , 'ー' ) __lowerCamelCase : List[Any] = text.replace('−' , 'ー' ) for k, v in self.emoji["emoji"].items(): if k in text: __lowerCamelCase : Optional[Any] = text.replace(a , a ) if clean: __lowerCamelCase : List[str] = self.clean_text(a ) def check_simbol(a: List[str] ): __lowerCamelCase : Dict = x.encode() if len(a ) == 1 and len(a ) == 2: __lowerCamelCase : Optional[Any] = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0Xc2a1 and c <= 0Xc2bf) or (c >= 0Xc780 and c <= 0Xc783) or (c >= 0Xcab9 and c <= 0Xcbbf) or (c >= 0Xcc80 and c <= 0Xcda2) ): return True return False def checkuae(a: Optional[int] ): __lowerCamelCase : Optional[Any] = x.encode() if len(a ) == 1 and len(a ) == 3: __lowerCamelCase : List[Any] = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0Xe2_8080 and c <= 0Xe2_b07f: return True return False __lowerCamelCase : Union[str, Any] = 0 __lowerCamelCase : str = [] while pos < len(a ): __lowerCamelCase : Optional[int] = min(len(a ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 __lowerCamelCase : Tuple = [] # (token_id, token, pos) for e in range(a , a , -1 ): __lowerCamelCase : Any = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(a ) > 2: __lowerCamelCase : int = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(a ) > 0: # the smallest token_id is adopted __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Union[str, Any] = sorted(a , key=lambda a : x[0] )[0] result.append(a ) __lowerCamelCase : Union[str, Any] = e else: __lowerCamelCase : Tuple = pos + 1 __lowerCamelCase : Dict = text[pos:end] if check_simbol(a ): result.append('<KIGOU>' ) elif checkuae(a ): result.append('<U2000U2BFF>' ) else: for i in wd.encode('utf-8' ): result.append('<|byte%d|>' % i ) __lowerCamelCase : str = end return result def _snake_case ( self: Optional[Any] , a: Optional[int] , a: str="\n" ): __lowerCamelCase : Dict = [] __lowerCamelCase : str = [] __lowerCamelCase : str = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(a ) > 0: words.append(bytearray(a ).decode('utf-8' , errors='replace' ) ) __lowerCamelCase : int = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['emoji_inv'][word] ) elif word == "<SP>": words.append(' ' ) elif word == "<BR>": words.append(a ) elif word == "<TAB>": words.append('\t' ) elif word == "<BLOCK>": words.append('▀' ) elif word == "<KIGOU>": words.append('ǀ' ) elif word == "<U2000U2BFF>": words.append('‖' ) else: words.append(a ) if len(a ) > 0: words.append(bytearray(a ).decode('utf-8' , errors='replace' ) ) __lowerCamelCase : List[Any] = ''.join(a ) return text
230
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class a : def __init__( self :List[Any] ,__lowercase :int ,__lowercase :Union[str, Any]=1_3 ,__lowercase :Optional[int]=7 ,__lowercase :Tuple=False ,__lowercase :Union[str, Any]=True ,__lowercase :Any=False ,__lowercase :Tuple=True ,__lowercase :str=3_3 ,__lowercase :Any=3_2 ,__lowercase :Any=5 ,__lowercase :int=4 ,__lowercase :Optional[Any]=3_7 ,__lowercase :List[str]="gelu" ,__lowercase :List[str]=0.1 ,__lowercase :Union[str, Any]=0.1 ,__lowercase :Dict=5_1_2 ,__lowercase :str=1_6 ,__lowercase :Dict=2 ,__lowercase :Optional[Any]=0.02 ,__lowercase :Tuple=3 ,__lowercase :Tuple=4 ,__lowercase :str=None ,): snake_case__ : Any = parent snake_case__ : Dict = batch_size snake_case__ : List[Any] = seq_length snake_case__ : List[str] = is_training snake_case__ : List[str] = use_input_mask snake_case__ : Union[str, Any] = use_token_type_ids snake_case__ : str = use_labels snake_case__ : Any = vocab_size snake_case__ : List[str] = hidden_size snake_case__ : str = num_hidden_layers snake_case__ : List[Any] = num_attention_heads snake_case__ : str = intermediate_size snake_case__ : List[str] = hidden_act snake_case__ : List[Any] = hidden_dropout_prob snake_case__ : Any = attention_probs_dropout_prob snake_case__ : Any = max_position_embeddings snake_case__ : Tuple = type_vocab_size snake_case__ : List[Any] = type_sequence_label_size snake_case__ : str = initializer_range snake_case__ : List[Any] = num_labels snake_case__ : Tuple = num_choices snake_case__ : int = scope def __lowerCamelCase ( self :List[str] ): snake_case__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) snake_case__ : str = None if self.use_input_mask: snake_case__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) snake_case__ : Optional[Any] = None snake_case__ : int = None snake_case__ : Any = None if self.use_labels: snake_case__ : Optional[int] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) snake_case__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) snake_case__ : Union[str, Any] = ids_tensor([self.batch_size] ,self.num_choices ) snake_case__ : Optional[int] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCamelCase ( self :Dict ): return EsmConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,pad_token_id=1 ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) def __lowerCamelCase ( self :Optional[int] ,__lowercase :int ,__lowercase :Optional[int] ,__lowercase :Dict ,__lowercase :Dict ,__lowercase :Optional[Any] ,__lowercase :Any ): snake_case__ : str = EsmModel(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case__ : Tuple = model(__lowercase ,attention_mask=__lowercase ) snake_case__ : int = model(__lowercase ) snake_case__ : Optional[int] = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def __lowerCamelCase ( self :Optional[int] ,__lowercase :Optional[int] ,__lowercase :int ,__lowercase :Optional[Any] ,__lowercase :Tuple ,__lowercase :int ,__lowercase :List[str] ): snake_case__ : List[str] = EsmForMaskedLM(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case__ : str = model(__lowercase ,attention_mask=__lowercase ,labels=__lowercase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCamelCase ( self :int ,__lowercase :int ,__lowercase :Tuple ,__lowercase :str ,__lowercase :Union[str, Any] ,__lowercase :List[str] ,__lowercase :Any ): snake_case__ : int = self.num_labels snake_case__ : Tuple = EsmForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case__ : Tuple = model(__lowercase ,attention_mask=__lowercase ,labels=__lowercase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def __lowerCamelCase ( self :str ): snake_case__ : Optional[Any] = self.prepare_config_and_inputs() ( ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ( snake_case__ ) , ) : Optional[Any] = config_and_inputs snake_case__ : Union[str, Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class a ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCAmelCase : List[str] = False __lowerCAmelCase : Any = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __lowerCAmelCase : Optional[int] = () __lowerCAmelCase : Optional[int] = ( { """feature-extraction""": EsmModel, """fill-mask""": EsmForMaskedLM, """text-classification""": EsmForSequenceClassification, """token-classification""": EsmForTokenClassification, """zero-shot""": EsmForSequenceClassification, } if is_torch_available() else {} ) __lowerCAmelCase : List[Any] = True def __lowerCamelCase ( self :Any ): snake_case__ : Union[str, Any] = EsmModelTester(self ) snake_case__ : str = ConfigTester(self ,config_class=__lowercase ,hidden_size=3_7 ) def __lowerCamelCase ( self :Optional[Any] ): self.config_tester.run_common_tests() def __lowerCamelCase ( self :Dict ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def __lowerCamelCase ( self :str ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case__ : Optional[Any] = type self.model_tester.create_and_check_model(*__lowercase ) def __lowerCamelCase ( self :Optional[int] ): snake_case__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowercase ) def __lowerCamelCase ( self :Any ): snake_case__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowercase ) @slow def __lowerCamelCase ( self :Any ): for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ : Union[str, Any] = EsmModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def __lowerCamelCase ( self :Optional[Any] ): snake_case__ : List[Any] = self.model_tester.prepare_config_and_inputs()[0] snake_case__ : List[Any] = EsmEmbeddings(config=__lowercase ) snake_case__ : int = torch.as_tensor([[1_2, 3_1, 1_3, model.padding_idx]] ) snake_case__ : Optional[int] = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) snake_case__ : Optional[Any] = create_position_ids_from_input_ids(__lowercase ,model.padding_idx ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(__lowercase ,__lowercase ) ) ) def __lowerCamelCase ( self :Union[str, Any] ): snake_case__ : str = self.model_tester.prepare_config_and_inputs()[0] snake_case__ : List[str] = EsmEmbeddings(config=__lowercase ) snake_case__ : Optional[int] = torch.empty(2 ,4 ,3_0 ) snake_case__ : Dict = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] snake_case__ : Union[str, Any] = torch.as_tensor([expected_single_positions, expected_single_positions] ) snake_case__ : Dict = embeddings.create_position_ids_from_inputs_embeds(__lowercase ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(__lowercase ,__lowercase ) ) ) @unittest.skip('''Esm does not support embedding resizing''' ) def __lowerCamelCase ( self :Optional[int] ): pass @unittest.skip('''Esm does not support embedding resizing''' ) def __lowerCamelCase ( self :Tuple ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __lowerCamelCase ( self :Union[str, Any] ): pass @require_torch class a ( __lowerCamelCase ): @slow def __lowerCamelCase ( self :int ): with torch.no_grad(): snake_case__ : str = EsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) model.eval() snake_case__ : List[Any] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) snake_case__ : str = model(__lowercase )[0] snake_case__ : str = 3_3 snake_case__ : Dict = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape ,__lowercase ) snake_case__ : str = torch.tensor( [[[8.9215, -10.5898, -6.4671], [-6.3967, -13.9114, -1.1212], [-7.7812, -13.9516, -3.7406]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,__lowercase ,atol=1e-4 ) ) @slow def __lowerCamelCase ( self :List[Any] ): with torch.no_grad(): snake_case__ : str = EsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) model.eval() snake_case__ : Union[str, Any] = torch.tensor([[0, 6, 4, 1_3, 5, 4, 1_6, 1_2, 1_1, 7, 2]] ) snake_case__ : str = model(__lowercase )[0] # compare the actual values for a slice. snake_case__ : List[Any] = torch.tensor( [[[0.1444, 0.5413, 0.3248], [0.3034, 0.0053, 0.3108], [0.3228, -0.2499, 0.3415]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,__lowercase ,atol=1e-4 ) )
252
def _lowerCAmelCase ( __lowerCAmelCase = 200 ) -> int: """simple docstring""" snake_case__ : Optional[int] = [1, 2, 5, 10, 20, 50, 100, 200] snake_case__ : List[Any] = [0] * (pence + 1) snake_case__ : str = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(__lowerCAmelCase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
252
1
'''simple docstring''' from __future__ import annotations A: int = list[list[int]] # assigning initial values to the grid A: Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution A: Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def _UpperCAmelCase ( a : Matrix , a : int , a : int , a : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def _UpperCAmelCase ( a : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def _UpperCAmelCase ( a : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(a ): lowercase_ , lowercase_ : List[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 1_0 ): if is_safe(a , a , a , a ): lowercase_ : str = digit if sudoku(a ) is not None: return grid lowercase_ : Union[str, Any] = 0 return None def _UpperCAmelCase ( a : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(a , end=' ' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("\nExample grid:\n" + "=" * 2_0) print_solution(example_grid) print("\nExample grid solution:") A: Any = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("Cannot find a solution.")
7
'''simple docstring''' from __future__ import annotations def _UpperCAmelCase ( a : int = 4 ) -> list[list[int]]: """simple docstring""" lowercase_ : Tuple = abs(a ) or 4 return [[1 + x + y * row_size for x in range(a )] for y in range(a )] def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" return reverse_row(transpose(a ) ) # OR.. transpose(reverse_column(matrix)) def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" return reverse_row(reverse_column(a ) ) # OR.. reverse_column(reverse_row(matrix)) def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" return reverse_column(transpose(a ) ) # OR.. transpose(reverse_row(matrix)) def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" lowercase_ : Any = [list(a ) for x in zip(*a )] return matrix def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" lowercase_ : List[str] = matrix[::-1] return matrix def _UpperCAmelCase ( a : list[list[int]] ) -> list[list[int]]: """simple docstring""" lowercase_ : str = [x[::-1] for x in matrix] return matrix def _UpperCAmelCase ( a : list[list[int]] ) -> None: """simple docstring""" for i in matrix: print(*a ) if __name__ == "__main__": A: Dict = make_matrix() print("\norigin:\n") print_matrix(matrix) print("\nrotate 90 counterclockwise:\n") print_matrix(rotate_aa(matrix)) A: List[Any] = make_matrix() print("\norigin:\n") print_matrix(matrix) print("\nrotate 180:\n") print_matrix(rotate_aaa(matrix)) A: List[str] = make_matrix() print("\norigin:\n") print_matrix(matrix) print("\nrotate 270 counterclockwise:\n") print_matrix(rotate_aaa(matrix))
7
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { 'configuration_mctct': ['MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MCTCTConfig'], 'feature_extraction_mctct': ['MCTCTFeatureExtractor'], 'processing_mctct': ['MCTCTProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ 'MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MCTCTForCTC', 'MCTCTModel', 'MCTCTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowercase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
562
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) lowercase_ = { 'iou_prediction_head.layers.0': 'iou_prediction_head.proj_in', 'iou_prediction_head.layers.1': 'iou_prediction_head.layers.0', 'iou_prediction_head.layers.2': 'iou_prediction_head.proj_out', 'mask_decoder.output_upscaling.0': 'mask_decoder.upscale_conv1', 'mask_decoder.output_upscaling.1': 'mask_decoder.upscale_layer_norm', 'mask_decoder.output_upscaling.3': 'mask_decoder.upscale_conv2', 'mask_downscaling.0': 'mask_embed.conv1', 'mask_downscaling.1': 'mask_embed.layer_norm1', 'mask_downscaling.3': 'mask_embed.conv2', 'mask_downscaling.4': 'mask_embed.layer_norm2', 'mask_downscaling.6': 'mask_embed.conv3', 'point_embeddings': 'point_embed', 'pe_layer.positional_encoding_gaussian_matrix': 'shared_embedding.positional_embedding', 'image_encoder': 'vision_encoder', 'neck.0': 'neck.conv1', 'neck.1': 'neck.layer_norm1', 'neck.2': 'neck.conv2', 'neck.3': 'neck.layer_norm2', 'patch_embed.proj': 'patch_embed.projection', '.norm': '.layer_norm', 'blocks': 'layers', } def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: _a = {} state_dict.pop('pixel_mean' , _UpperCAmelCase ) state_dict.pop('pixel_std' , _UpperCAmelCase ) _a = R'.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _a = key.replace(_UpperCAmelCase , _UpperCAmelCase ) if re.match(_UpperCAmelCase , _UpperCAmelCase ): _a = int(re.match(_UpperCAmelCase , _UpperCAmelCase ).group(2 ) ) if layer_nb == 0: _a = key.replace('layers.0' , 'proj_in' ) elif layer_nb == 1: _a = key.replace('layers.1' , 'layers.0' ) elif layer_nb == 2: _a = key.replace('layers.2' , 'proj_out' ) _a = value _a = model_state_dict[ 'prompt_encoder.shared_embedding.positional_embedding' ] return model_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase="ybelkada/segment-anything" ) -> Optional[Any]: _a = hf_hub_download(_UpperCAmelCase , f"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: _a = SamConfig() elif "sam_vit_l" in model_name: _a = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) _a = SamConfig( vision_config=_UpperCAmelCase , ) elif "sam_vit_h" in model_name: _a = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) _a = SamConfig( vision_config=_UpperCAmelCase , ) _a = torch.load(_UpperCAmelCase , map_location='cpu' ) _a = replace_keys(_UpperCAmelCase ) _a = SamImageProcessor() _a = SamProcessor(image_processor=_UpperCAmelCase ) _a = SamModel(_UpperCAmelCase ) hf_model.load_state_dict(_UpperCAmelCase ) _a = hf_model.to('cuda' ) _a = 'https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png' _a = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ).convert('RGB' ) _a = [[[400, 650]]] _a = [[1]] _a = processor(images=np.array(_UpperCAmelCase ) , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): _a = hf_model(**_UpperCAmelCase ) _a = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 _a = processor( images=np.array(_UpperCAmelCase ) , input_points=_UpperCAmelCase , input_labels=_UpperCAmelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): _a = hf_model(**_UpperCAmelCase ) _a = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 _a = ((75, 275, 1725, 850),) _a = processor(images=np.array(_UpperCAmelCase ) , input_boxes=_UpperCAmelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): _a = hf_model(**_UpperCAmelCase ) _a = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. _a = [[[400, 650], [800, 650]]] _a = [[1, 1]] _a = processor( images=np.array(_UpperCAmelCase ) , input_points=_UpperCAmelCase , input_labels=_UpperCAmelCase , return_tensors='pt' ).to('cuda' ) with torch.no_grad(): _a = hf_model(**_UpperCAmelCase ) _a = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() lowercase_ = ['sam_vit_b_01ec64', 'sam_vit_h_4b8939', 'sam_vit_l_0b3195'] parser.add_argument( '--model_name', default='sam_vit_h_4b8939', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) parser.add_argument( '--model_hub_id', default='ybelkada/segment-anything', choices=choices, type=str, help='Path to hf config.json of model to convert', ) lowercase_ = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
562
1
'''simple docstring''' import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class a_ ( lowerCAmelCase__ ): lowercase = 42 lowercase = 42 class a_ ( lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = 1 @register_to_config def __init__( self , _SCREAMING_SNAKE_CASE = 2000 , _SCREAMING_SNAKE_CASE = 0.1_5 , _SCREAMING_SNAKE_CASE = 0.0_1 , _SCREAMING_SNAKE_CASE = 1348.0 , _SCREAMING_SNAKE_CASE = 1e-5 , _SCREAMING_SNAKE_CASE = 1 , ) -> List[str]: """simple docstring""" UpperCamelCase = sigma_max # setable values UpperCamelCase = None self.set_sigmas(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> torch.FloatTensor: """simple docstring""" return sample def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None ) -> List[str]: """simple docstring""" UpperCamelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps UpperCamelCase = torch.linspace(1 , _lowerCamelCase , _lowerCamelCase , device=_lowerCamelCase ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None ) -> List[Any]: """simple docstring""" UpperCamelCase = sigma_min if sigma_min is not None else self.config.sigma_min UpperCamelCase = sigma_max if sigma_max is not None else self.config.sigma_max UpperCamelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(_lowerCamelCase , _lowerCamelCase ) UpperCamelCase = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) UpperCamelCase = torch.exp(torch.linspace(math.log(_lowerCamelCase ) , math.log(_lowerCamelCase ) , _lowerCamelCase ) ) UpperCamelCase = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = True , ) -> Union[SdeVeOutput, Tuple]: """simple docstring""" if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler""" ) UpperCamelCase = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) UpperCamelCase = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda UpperCamelCase = timesteps.to(self.discrete_sigmas.device ) UpperCamelCase = self.discrete_sigmas[timesteps].to(sample.device ) UpperCamelCase = self.get_adjacent_sigma(_lowerCamelCase , _lowerCamelCase ).to(sample.device ) UpperCamelCase = torch.zeros_like(_lowerCamelCase ) UpperCamelCase = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods UpperCamelCase = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): UpperCamelCase = diffusion.unsqueeze(-1 ) UpperCamelCase = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of UpperCamelCase = randn_tensor( sample.shape , layout=sample.layout , generator=_lowerCamelCase , device=sample.device , dtype=sample.dtype ) UpperCamelCase = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? UpperCamelCase = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=_lowerCamelCase , prev_sample_mean=_lowerCamelCase ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler""" ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction UpperCamelCase = randn_tensor(sample.shape , layout=sample.layout , generator=_lowerCamelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr UpperCamelCase = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() UpperCamelCase = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() UpperCamelCase = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 UpperCamelCase = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term UpperCamelCase = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): UpperCamelCase = step_size.unsqueeze(-1 ) UpperCamelCase = sample + step_size * model_output UpperCamelCase = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCamelCase ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> torch.FloatTensor: """simple docstring""" UpperCamelCase = timesteps.to(original_samples.device ) UpperCamelCase = self.discrete_sigmas.to(original_samples.device )[timesteps] UpperCamelCase = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(_lowerCamelCase ) * sigmas[:, None, None, None] ) UpperCamelCase = noise + original_samples return noisy_samples def __len__( self ) -> Union[str, Any]: """simple docstring""" return self.config.num_train_timesteps
711
'''simple docstring''' import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class a_ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=[1, 2, 1] , _SCREAMING_SNAKE_CASE=[2, 2, 4] , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=2.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=["stage1", "stage2", "stage3"] , _SCREAMING_SNAKE_CASE=[1, 2, 3] , ) -> Any: """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = embed_dim UpperCamelCase = depths UpperCamelCase = num_heads UpperCamelCase = window_size UpperCamelCase = mlp_ratio UpperCamelCase = qkv_bias UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = drop_path_rate UpperCamelCase = hidden_act UpperCamelCase = use_absolute_embeddings UpperCamelCase = patch_norm UpperCamelCase = layer_norm_eps UpperCamelCase = initializer_range UpperCamelCase = is_training UpperCamelCase = scope UpperCamelCase = use_labels UpperCamelCase = type_sequence_label_size UpperCamelCase = encoder_stride UpperCamelCase = out_features UpperCamelCase = out_indices def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase = self.get_config() return config, pixel_values, labels def A__ ( self ) -> str: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = MaskFormerSwinModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) UpperCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" UpperCamelCase = MaskFormerSwinBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_SCREAMING_SNAKE_CASE ): UpperCamelCase = ["""stem"""] UpperCamelCase = MaskFormerSwinBackbone(config=_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase ,UpperCamelCase ,UpperCamelCase = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a_ ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase = {"""feature-extraction""": MaskFormerSwinModel} if is_torch_available() else {} lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False def A__ ( self ) -> int: """simple docstring""" UpperCamelCase = MaskFormerSwinModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def A__ ( self ) -> List[str]: """simple docstring""" pass def A__ ( self ) -> Dict: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self ) -> int: """simple docstring""" return def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_SCREAMING_SNAKE_CASE ) @unittest.skip("""Swin does not use inputs_embeds""" ) def A__ ( self ) -> Optional[Any]: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def A__ ( self ) -> Dict: """simple docstring""" pass def A__ ( self ) -> Optional[int]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear ) ) def A__ ( self ) -> int: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def A__ ( self ) -> Optional[int]: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def A__ ( self ) -> List[str]: """simple docstring""" pass def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCamelCase = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase = outputs.hidden_states UpperCamelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # Swin has a different seq_length UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def A__ ( self ) -> str: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = 3 UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def A__ ( self ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def A__ ( self ) -> List[Any]: """simple docstring""" pass def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_SCREAMING_SNAKE_CASE ): UpperCamelCase = 0 return t def check_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE={} ): with torch.no_grad(): UpperCamelCase = model(**_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) UpperCamelCase = model(**_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ).to_tuple() def recursive_check(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if isinstance(_SCREAMING_SNAKE_CASE , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): recursive_check(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_SCREAMING_SNAKE_CASE ) , set_nan_tensor_to_zero(_SCREAMING_SNAKE_CASE ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(_SCREAMING_SNAKE_CASE ).any()} and `inf`: {torch.isinf(_SCREAMING_SNAKE_CASE )}. Dict has" F" `nan`: {torch.isnan(_SCREAMING_SNAKE_CASE ).any()} and `inf`: {torch.isinf(_SCREAMING_SNAKE_CASE )}." ) , ) recursive_check(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) check_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) check_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) check_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , {"""output_hidden_states""": True} ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) UpperCamelCase = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) check_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , {"""output_hidden_states""": True} ) @require_torch class a_ ( unittest.TestCase , lowerCamelCase ): lowercase = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase = MaskFormerSwinConfig def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = MaskFormerSwinModelTester(self ) def A__ ( self ) -> str: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: UpperCamelCase = backbone_class(_SCREAMING_SNAKE_CASE ) backbone.to(_SCREAMING_SNAKE_CASE ) backbone.eval() UpperCamelCase = backbone(**_SCREAMING_SNAKE_CASE ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _SCREAMING_SNAKE_CASE ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True UpperCamelCase = backbone(**_SCREAMING_SNAKE_CASE , output_hidden_states=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) UpperCamelCase ,UpperCamelCase ,UpperCamelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: UpperCamelCase = backbone(**_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(outputs.attentions )
35
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case : Union[str, Any] = logging.get_logger(__name__) _snake_case : Tuple = {"vocab_file": "spm_char.model"} _snake_case : str = { "vocab_file": { "microsoft/speecht5_asr": "https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model", "microsoft/speecht5_tts": "https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model", "microsoft/speecht5_vc": "https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model", } } _snake_case : Tuple = { "microsoft/speecht5_asr": 1_024, "microsoft/speecht5_tts": 1_024, "microsoft/speecht5_vc": 1_024, } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : str="<s>" , lowerCamelCase : Union[str, Any]="</s>" , lowerCamelCase : Any="<unk>" , lowerCamelCase : Union[str, Any]="<pad>" , lowerCamelCase : Optional[Dict[str, Any]] = None , **lowerCamelCase : Any , ) -> None: __snake_case : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase , eos_token=lowerCamelCase , unk_token=lowerCamelCase , pad_token=lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase , ) __snake_case : Dict = vocab_file __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase ) @property def __snake_case ( self : int ) -> Tuple: return self.sp_model.get_piece_size() def __snake_case ( self : List[str] ) -> Any: __snake_case : List[str] = {self.convert_ids_to_tokens(lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ) -> Union[str, Any]: __snake_case : Optional[int] = self.__dict__.copy() __snake_case : Optional[Any] = None return state def __setstate__( self : int , lowerCamelCase : str ) -> Tuple: __snake_case : Tuple = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __snake_case : Tuple = {} __snake_case : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __snake_case ( self : List[str] , lowerCamelCase : str ) -> List[str]: return self.sp_model.encode(lowerCamelCase , out_type=lowerCamelCase ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Union[str, Any] ) -> Optional[int]: return self.sp_model.piece_to_id(lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : Dict ) -> Tuple: __snake_case : List[Any] = self.sp_model.IdToPiece(lowerCamelCase ) return token def __snake_case ( self : int , lowerCamelCase : Any ) -> Tuple: __snake_case : Optional[Any] = [] __snake_case : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCamelCase ) + token __snake_case : Tuple = [] else: current_sub_tokens.append(lowerCamelCase ) out_string += self.sp_model.decode(lowerCamelCase ) return out_string.strip() def __snake_case ( self : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any]=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def __snake_case ( self : Optional[Any] , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None , lowerCamelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase , token_ids_a=lowerCamelCase , already_has_special_tokens=lowerCamelCase ) __snake_case : str = [1] if token_ids_a is None: return ([0] * len(lowerCamelCase )) + suffix_ones return ([0] * len(lowerCamelCase )) + ([0] * len(lowerCamelCase )) + suffix_ones def __snake_case ( self : Any , lowerCamelCase : str , lowerCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCamelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __snake_case : Optional[int] = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase , "wb" ) as fi: __snake_case : int = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase ) return (out_vocab_file,)
81
'''simple docstring''' from __future__ import annotations import os from collections.abc import Mapping _UpperCamelCase : Any = tuple[int, int] class snake_case__ : def __init__( self : List[str] , _A : set[int] , _A : Mapping[EdgeT, int] ) -> None: UpperCAmelCase_ : set[int] = vertices UpperCAmelCase_ : dict[EdgeT, int] = { (min(_A ), max(_A )): weight for edge, weight in edges.items() } def A ( self : Union[str, Any] , _A : EdgeT , _A : int ) -> None: self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) UpperCAmelCase_ : List[str] = weight def A ( self : str ) -> Graph: UpperCAmelCase_ : Graph = Graph({min(self.vertices )} , {} ) UpperCAmelCase_ : EdgeT UpperCAmelCase_ : int UpperCAmelCase_ : EdgeT UpperCAmelCase_ : int while len(subgraph.vertices ) < len(self.vertices ): UpperCAmelCase_ : Optional[Any] = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: UpperCAmelCase_ : Dict = edge UpperCAmelCase_ : Tuple = weight subgraph.add_edge(_A , _A ) return subgraph def __UpperCAmelCase ( A : str = "p107_network.txt" ) -> int: UpperCAmelCase_ : str = os.path.abspath(os.path.dirname(A ) ) UpperCAmelCase_ : str = os.path.join(A , A ) UpperCAmelCase_ : dict[EdgeT, int] = {} UpperCAmelCase_ : list[str] UpperCAmelCase_ : int UpperCAmelCase_ : int with open(A ) as f: UpperCAmelCase_ : int = f.read().strip().split('''\n''' ) UpperCAmelCase_ : Any = [line.split(''',''' ) for line in data] for edgea in range(1 , len(A ) ): for edgea in range(A ): if adjaceny_matrix[edgea][edgea] != "-": UpperCAmelCase_ : Optional[int] = int(adjaceny_matrix[edgea][edgea] ) UpperCAmelCase_ : Graph = Graph(set(range(len(A ) ) ) , A ) UpperCAmelCase_ : Graph = graph.prims_algorithm() UpperCAmelCase_ : int = sum(graph.edges.values() ) UpperCAmelCase_ : int = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(f'''{solution() = }''')
541
0
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def A__ ( UpperCamelCase__="ro" , UpperCamelCase__="en" , UpperCamelCase__="wmt16" , UpperCamelCase__=None ): '''simple docstring''' try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) _SCREAMING_SNAKE_CASE = F'''{src_lang}-{tgt_lang}''' print(F'''Converting {dataset}-{pair}''' ) _SCREAMING_SNAKE_CASE = datasets.load_dataset(_snake_case , _snake_case ) if save_dir is None: _SCREAMING_SNAKE_CASE = F'''{dataset}-{pair}''' _SCREAMING_SNAKE_CASE = Path(_snake_case ) save_dir.mkdir(exist_ok=_snake_case ) for split in ds.keys(): print(F'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets _SCREAMING_SNAKE_CASE = '''val''' if split == '''validation''' else split _SCREAMING_SNAKE_CASE = save_dir.joinpath(F'''{fn}.source''' ) _SCREAMING_SNAKE_CASE = save_dir.joinpath(F'''{fn}.target''' ) _SCREAMING_SNAKE_CASE = src_path.open('''w+''' ) _SCREAMING_SNAKE_CASE = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): _SCREAMING_SNAKE_CASE = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(F'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
708
"""simple docstring""" from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class __snake_case( __A ): def __lt__( self , A_ ): '''simple docstring''' return self[-1] < other[-1] def __eq__( self , A_ ): '''simple docstring''' return self[-1] == other[-1] def A__ ( UpperCamelCase__ ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] # sort into stacks for element in collection: _SCREAMING_SNAKE_CASE = Stack([element] ) _SCREAMING_SNAKE_CASE = bisect_left(UpperCamelCase__ , UpperCamelCase__ ) if i != len(UpperCamelCase__ ): stacks[i].append(UpperCamelCase__ ) else: stacks.append(UpperCamelCase__ ) # use a heap-based merge to merge stack efficiently _SCREAMING_SNAKE_CASE = merge(*(reversed(UpperCamelCase__ ) for stack in stacks) ) return collection if __name__ == "__main__": lowerCamelCase : int = input("""Enter numbers separated by a comma:\n""").strip() lowerCamelCase : Optional[Any] = [int(item) for item in user_input.split(""",""")] print(patience_sort(unsorted))
168
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : List[Any] = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "swinv2" a = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : str , __lowerCamelCase : Optional[int]=224 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : Dict=3 , __lowerCamelCase : Any=96 , __lowerCamelCase : Optional[Any]=[2, 2, 6, 2] , __lowerCamelCase : Dict=[3, 6, 12, 24] , __lowerCamelCase : List[str]=7 , __lowerCamelCase : Tuple=4.0 , __lowerCamelCase : str=True , __lowerCamelCase : str=0.0 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Union[str, Any]=False , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Optional[int]=1e-5 , __lowerCamelCase : Optional[Any]=32 , **__lowerCamelCase : Optional[Any] , ) -> List[Any]: super().__init__(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = len(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = window_size SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = use_absolute_embeddings SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ = int(embed_dim * 2 ** (len(__lowerCamelCase ) - 1) ) SCREAMING_SNAKE_CASE__ = (0, 0, 0, 0)
493
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 _SCREAMING_SNAKE_CASE : Optional[Any] = get_tests_dir('''fixtures''') class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self : Any ) -> Any: # A mock response for an HTTP head request to emulate server down SCREAMING_SNAKE_CASE__ = mock.Mock() SCREAMING_SNAKE_CASE__ = 500 SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = HTTPError SCREAMING_SNAKE_CASE__ = {} # Download this model to make sure it's in the cache. SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained('''hf-internal-testing/tiny-random-wav2vec2''' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=__lowerCamelCase ) as mock_head: SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained('''hf-internal-testing/tiny-random-wav2vec2''' ) # This check we did call the fake head request mock_head.assert_called() def lowercase_ ( self : int ) -> Dict: # This test is for deprecated behavior and can be removed in v5 SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json''' ) @is_staging_test class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @classmethod def lowercase_ ( cls : Union[str, Any] ) -> List[str]: SCREAMING_SNAKE_CASE__ = TOKEN HfFolder.save_token(__lowerCamelCase ) @classmethod def lowercase_ ( cls : Optional[int] ) -> Optional[int]: try: delete_repo(token=cls._token , repo_id='''test-feature-extractor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-feature-extractor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-feature-extractor''' ) except HTTPError: pass def lowercase_ ( self : Optional[Any] ) -> int: SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(__lowerCamelCase ) feature_extractor.push_to_hub('''test-feature-extractor''' , use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(f'''{USER}/test-feature-extractor''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(__lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-feature-extractor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( __lowerCamelCase , repo_id='''test-feature-extractor''' , push_to_hub=__lowerCamelCase , use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(f'''{USER}/test-feature-extractor''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(__lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) def lowercase_ ( self : Tuple ) -> Any: SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(__lowerCamelCase ) feature_extractor.push_to_hub('''valid_org/test-feature-extractor''' , use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained('''valid_org/test-feature-extractor''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(__lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-feature-extractor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( __lowerCamelCase , repo_id='''valid_org/test-feature-extractor-org''' , push_to_hub=__lowerCamelCase , use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained('''valid_org/test-feature-extractor-org''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(__lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) def lowercase_ ( self : int ) -> int: CustomFeatureExtractor.register_for_auto_class() SCREAMING_SNAKE_CASE__ = CustomFeatureExtractor.from_pretrained(__lowerCamelCase ) feature_extractor.push_to_hub('''test-dynamic-feature-extractor''' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {'''AutoFeatureExtractor''': '''custom_feature_extraction.CustomFeatureExtractor'''} , ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( f'''{USER}/test-dynamic-feature-extractor''' , trust_remote_code=__lowerCamelCase ) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , '''CustomFeatureExtractor''' )
493
1
from typing import List import numpy as np def lowerCamelCase__ ( _lowercase ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = {key: len(_lowercase ) for key, value in gen_kwargs.items() if isinstance(_lowercase , _lowercase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(f'''\t- key {key} has length {length}''' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) UpperCAmelCase_ : Tuple = max(lists_lengths.values() , default=0 ) return max(1 , _lowercase ) def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = [] for group_idx in range(_lowercase ): UpperCAmelCase_ : Any = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break UpperCAmelCase_ : int = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 UpperCAmelCase_ : Tuple = range(_lowercase , start + num_shards_to_add ) shards_indices_per_group.append(_lowercase ) return shards_indices_per_group def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Dict = _number_of_shards_in_gen_kwargs(_lowercase ) if num_shards == 1: return [dict(_lowercase )] else: UpperCAmelCase_ : str = _distribute_shards(num_shards=_lowercase , max_num_jobs=_lowercase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(_lowercase , _lowercase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(_lowercase ) ) ] def lowerCamelCase__ ( _lowercase ): '''simple docstring''' return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , _lowercase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = {len(_lowercase ) for value in gen_kwargs.values() if isinstance(_lowercase , _lowercase )} UpperCAmelCase_ : List[str] = {} for size in list_sizes: UpperCAmelCase_ : Optional[Any] = list(range(_lowercase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes UpperCAmelCase_ : Optional[int] = dict(_lowercase ) for key, value in shuffled_kwargs.items(): if isinstance(_lowercase , _lowercase ): UpperCAmelCase_ : List[Any] = [value[i] for i in indices_per_size[len(_lowercase )]] return shuffled_kwargs
300
def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' if digit_amount > 0: return round(number - int(_lowercase ) , _lowercase ) return number - int(_lowercase ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
300
1
import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase__ : List[str] = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right lowerCamelCase__ : int = 50_003 lowerCamelCase__ : List[str] = 50_002 @require_sentencepiece @require_tokenizers class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = PLBartTokenizer lowercase_ = None lowercase_ = False def lowerCAmelCase_ ( self : List[str] ): super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE_ = PLBartTokenizer(_lowerCAmelCase , language_codes='base' , keep_accents=_lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = PLBartTokenizer(_lowerCAmelCase , language_codes='base' , keep_accents=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) SCREAMING_SNAKE_CASE_ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.vocab_size SCREAMING_SNAKE_CASE_ = [tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) for x in range(end - 4 , _lowerCAmelCase )] self.assertListEqual(_lowerCAmelCase , ['__java__', '__python__', '__en_XX__', '<mask>'] ) SCREAMING_SNAKE_CASE_ = 'java.lang.Exception, python.lang.Exception, javascript, php, ruby, go' SCREAMING_SNAKE_CASE_ = tokenizer(_lowerCAmelCase ).input_ids self.assertEqual( tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) , _lowerCAmelCase , ) def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = PLBartTokenizer(_lowerCAmelCase , language_codes='multi' , keep_accents=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) SCREAMING_SNAKE_CASE_ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) SCREAMING_SNAKE_CASE_ = tokenizer.vocab_size SCREAMING_SNAKE_CASE_ = [tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) for x in range(end - 7 , _lowerCAmelCase )] self.assertListEqual( _lowerCAmelCase , ['__java__', '__python__', '__en_XX__', '__javascript__', '__php__', '__ruby__', '__go__'] ) SCREAMING_SNAKE_CASE_ = 'java.lang.Exception, python.lang.Exception, javascript, php, ruby, go' SCREAMING_SNAKE_CASE_ = tokenizer(_lowerCAmelCase ).input_ids self.assertEqual( tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) , _lowerCAmelCase , ) @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' lowercase_ = "uclanlp/plbart-python-en_XX" lowercase_ = [ "def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])", "def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])", ] lowercase_ = [ "Returns the maximum value of a b c.", "Sums the values of a b c.", ] lowercase_ = [ 134, 5_452, 33_460, 33_441, 33_463, 33_465, 33_463, 33_449, 988, 20, 33_456, 19, 33_456, 771, 39, 4_258, 889, 3_318, 33_441, 33_463, 33_465, 33_463, 33_449, 2_471, 2, PYTHON_CODE, ] @classmethod def lowerCAmelCase_ ( cls : Optional[int] ): SCREAMING_SNAKE_CASE_ = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes='base' , src_lang='python' , tgt_lang='en_XX' ) SCREAMING_SNAKE_CASE_ = 1 return cls def lowerCAmelCase_ ( self : Union[str, Any] ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__java__'] , 50_001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__python__'] , 50_002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__en_XX__'] , 50_003 ) def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Any ): self.assertIn(_lowerCAmelCase , self.tokenizer.all_special_ids ) SCREAMING_SNAKE_CASE_ = [EN_CODE, 9_037, 33_442, 57, 752, 153, 14, 56, 18, 9, 2] SCREAMING_SNAKE_CASE_ = self.tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertNotIn(self.tokenizer.eos_token , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = ['def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])' * 20] self.assertIsInstance(src_text[0] , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = 10 SCREAMING_SNAKE_CASE_ = self.tokenizer(_lowerCAmelCase , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , _lowerCAmelCase ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', '__java__'] ) , [50_004, 50_001] ) def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = PLBartTokenizer.from_pretrained(_lowerCAmelCase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _lowerCAmelCase ) @require_torch def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , return_tensors='pt' ) SCREAMING_SNAKE_CASE_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , _lowerCAmelCase ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) SCREAMING_SNAKE_CASE_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 26) , batch.input_ids.shape ) self.assertEqual((2, 26) , batch.attention_mask.shape ) SCREAMING_SNAKE_CASE_ = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = self.tokenizer(self.src_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=3 , return_tensors='pt' ) SCREAMING_SNAKE_CASE_ = self.tokenizer( text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=10 , return_tensors='pt' ) SCREAMING_SNAKE_CASE_ = targets['input_ids'] SCREAMING_SNAKE_CASE_ = shift_tokens_right(_lowerCAmelCase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='java' ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , { # A, test, EOS, en_XX 'input_ids': [[150, 242, 2, 50_003]], 'attention_mask': [[1, 1, 1, 1]], # java 'forced_bos_token_id': 50_001, } , )
31
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( 'files' , [ ['full:README.md', 'dataset_infos.json'], ['empty:README.md', 'dataset_infos.json'], ['dataset_infos.json'], ['full:README.md'], ] , ) def a_ ( _A , _A ) -> List[Any]: """simple docstring""" snake_case__ = tmp_path_factory.mktemp('dset_infos_dir' ) if "full:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('---\ndataset_info:\n dataset_size: 42\n---' ) if "empty:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / 'dataset_infos.json' , 'w' ) as f: f.write('{"default": {"dataset_size": 42}}' ) snake_case__ = DatasetInfosDict.from_directory(_A ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( 'dataset_info' , [ DatasetInfo(), DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ), ] , ) def a_ ( _A , _A ) -> Optional[int]: """simple docstring""" snake_case__ = str(_A ) dataset_info.write_to_directory(_A ) snake_case__ = DatasetInfo.from_directory(_A ) assert dataset_info == reloaded assert os.path.exists(os.path.join(_A , 'dataset_info.json' ) ) def a_ ( ) -> Optional[int]: """simple docstring""" snake_case__ = DatasetInfo( description='foo' , citation='bar' , homepage='https://foo.bar' , license='CC0' , features=Features({'a': Value('int32' )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train', 'num_examples': 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) snake_case__ = dataset_info._to_yaml_dict() assert sorted(_A ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) snake_case__ = yaml.safe_dump(_A ) snake_case__ = yaml.safe_load(_A ) assert dataset_info_yaml_dict == reloaded def a_ ( ) -> Optional[Any]: """simple docstring""" snake_case__ = DatasetInfo() snake_case__ = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( 'dataset_infos_dict' , [ DatasetInfosDict(), DatasetInfosDict({'default': DatasetInfo()} ), DatasetInfosDict({'my_config_name': DatasetInfo()} ), DatasetInfosDict( { 'default': DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ) } ), DatasetInfosDict( { 'v1': DatasetInfo(dataset_size=42 ), 'v2': DatasetInfo(dataset_size=1337 ), } ), ] , ) def a_ ( _A , _A ) -> str: """simple docstring""" snake_case__ = str(_A ) dataset_infos_dict.write_to_directory(_A ) snake_case__ = DatasetInfosDict.from_directory(_A ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): snake_case__ = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml snake_case__ = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(_A , 'README.md' ) )
328
0
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __magic_name__ ( A_ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase = DanceDiffusionPipeline __UpperCamelCase = UNCONDITIONAL_AUDIO_GENERATION_PARAMS __UpperCamelCase = PipelineTesterMixin.required_optional_params - { '''callback''', '''latents''', '''callback_steps''', '''output_type''', '''num_images_per_prompt''', } __UpperCamelCase = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS __UpperCamelCase = False __UpperCamelCase = False def _lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCamelCase = UNetaDModel( block_out_channels=(32, 32, 64) , extra_in_channels=16 , sample_size=512 , sample_rate=16_000 , in_channels=2 , out_channels=2 , flip_sin_to_cos=_a , use_timestep_embedding=_a , time_embedding_type="""fourier""" , mid_block_type="""UNetMidBlock1D""" , down_block_types=("""DownBlock1DNoSkip""", """DownBlock1D""", """AttnDownBlock1D""") , up_block_types=("""AttnUpBlock1D""", """UpBlock1D""", """UpBlock1DNoSkip""") , ) lowerCamelCase = IPNDMScheduler() lowerCamelCase = { """unet""": unet, """scheduler""": scheduler, } return components def _lowerCAmelCase ( self , _a , _a=0 ): """simple docstring""" if str(_a ).startswith("""mps""" ): lowerCamelCase = torch.manual_seed(_a ) else: lowerCamelCase = torch.Generator(device=_a ).manual_seed(_a ) lowerCamelCase = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 4, } return inputs def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.get_dummy_components() lowerCamelCase = DanceDiffusionPipeline(**_a ) lowerCamelCase = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCamelCase = self.get_dummy_inputs(_a ) lowerCamelCase = pipe(**_a ) lowerCamelCase = output.audios lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) lowerCamelCase = np.array([-0.7_265, 1.0_000, -0.8_388, 0.1_175, 0.9_498, -1.0_000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def _lowerCAmelCase ( self ): """simple docstring""" return super().test_save_load_local() @skip_mps def _lowerCAmelCase ( self ): """simple docstring""" return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def _lowerCAmelCase ( self ): """simple docstring""" return super().test_save_load_optional_components() @skip_mps def _lowerCAmelCase ( self ): """simple docstring""" return super().test_attention_slicing_forward_pass() def _lowerCAmelCase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = torch_device lowerCamelCase = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" ) lowerCamelCase = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = pipe(generator=_a , num_inference_steps=100 , audio_length_in_s=4.096 ) lowerCamelCase = output.audios lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) lowerCamelCase = np.array([-0.0_192, -0.0_231, -0.0_318, -0.0_059, 0.0_002, -0.0_020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = torch_device lowerCamelCase = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" , torch_dtype=torch.floataa ) lowerCamelCase = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = pipe(generator=_a , num_inference_steps=100 , audio_length_in_s=4.096 ) lowerCamelCase = output.audios lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) lowerCamelCase = np.array([-0.0_367, -0.0_488, -0.0_771, -0.0_525, -0.0_444, -0.0_341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
708
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : List[str] = { """configuration_autoformer""": [ """AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AutoformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Tuple = [ """AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """AutoformerForPrediction""", """AutoformerModel""", """AutoformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys lowerCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
533
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ : int = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Any = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys lowercase__ : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
376
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowerCAmelCase_ = (DPMSolverSinglestepScheduler,) lowerCAmelCase_ = (('''num_inference_steps''', 25),) def snake_case__ ( self : str , **__lowercase : Any ): """simple docstring""" snake_case_ = { "num_train_timesteps": 10_00, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", "solver_order": 2, "prediction_type": "epsilon", "thresholding": False, "sample_max_value": 1.0, "algorithm_type": "dpmsolver++", "solver_type": "midpoint", "lambda_min_clipped": -float("inf" ), "variance_type": None, } config.update(**__lowercase ) return config def snake_case__ ( self : str , __lowercase : int=0 , **__lowercase : str ): """simple docstring""" snake_case_ = dict(self.forward_default_kwargs ) snake_case_ = kwargs.pop("num_inference_steps" , __lowercase ) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**__lowercase ) snake_case_ = scheduler_class(**__lowercase ) scheduler.set_timesteps(__lowercase ) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__lowercase ) snake_case_ = scheduler_class.from_pretrained(__lowercase ) new_scheduler.set_timesteps(__lowercase ) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(__lowercase , time_step + scheduler.config.solver_order + 1 ): snake_case_ = scheduler.step(__lowercase , __lowercase , __lowercase , **__lowercase ).prev_sample snake_case_ = new_scheduler.step(__lowercase , __lowercase , __lowercase , **__lowercase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def snake_case__ ( self : Union[str, Any] ): """simple docstring""" pass def snake_case__ ( self : List[Any] , __lowercase : Optional[int]=0 , **__lowercase : int ): """simple docstring""" snake_case_ = dict(self.forward_default_kwargs ) snake_case_ = kwargs.pop("num_inference_steps" , __lowercase ) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**__lowercase ) scheduler.set_timesteps(__lowercase ) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__lowercase ) snake_case_ = scheduler_class.from_pretrained(__lowercase ) # copy over dummy past residuals new_scheduler.set_timesteps(__lowercase ) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(__lowercase , __lowercase , __lowercase , **__lowercase ).prev_sample snake_case_ = new_scheduler.step(__lowercase , __lowercase , __lowercase , **__lowercase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def snake_case__ ( self : List[str] , __lowercase : Any=None , **__lowercase : List[Any] ): """simple docstring""" if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**__lowercase ) snake_case_ = scheduler_class(**__lowercase ) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**__lowercase ) snake_case_ = scheduler_class(**__lowercase ) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(__lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case_ = model(__lowercase , __lowercase ) snake_case_ = scheduler.step(__lowercase , __lowercase , __lowercase ).prev_sample return sample def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(__lowercase ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): snake_case_ = model(__lowercase , __lowercase ) snake_case_ = scheduler.step(__lowercase , __lowercase , __lowercase ).prev_sample snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.2574 ) < 1E-3 def snake_case__ ( self : Optional[int] ): """simple docstring""" for timesteps in [25, 50, 1_00, 9_99, 10_00]: self.check_over_configs(num_train_timesteps=__lowercase ) def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) snake_case_ = self.full_loop(scheduler=__lowercase ) snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.2791 ) < 1E-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config ) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config ) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config ) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config ) snake_case_ = self.full_loop(scheduler=__lowercase ) snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.2791 ) < 1E-3 def snake_case__ ( self : List[str] ): """simple docstring""" self.check_over_configs(thresholding=__lowercase ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=__lowercase , prediction_type=__lowercase , sample_max_value=__lowercase , algorithm_type="dpmsolver++" , solver_order=__lowercase , solver_type=__lowercase , ) def snake_case__ ( self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__lowercase ) def snake_case__ ( self : Optional[Any] ): """simple docstring""" for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=__lowercase , solver_type=__lowercase , prediction_type=__lowercase , algorithm_type=__lowercase , ) snake_case_ = self.full_loop( solver_order=__lowercase , solver_type=__lowercase , prediction_type=__lowercase , algorithm_type=__lowercase , ) assert not torch.isnan(__lowercase ).any(), "Samples have nan numbers" def snake_case__ ( self : List[Any] ): """simple docstring""" self.check_over_configs(lower_order_final=__lowercase ) self.check_over_configs(lower_order_final=__lowercase ) def snake_case__ ( self : Union[str, Any] ): """simple docstring""" self.check_over_configs(lambda_min_clipped=-float("inf" ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def snake_case__ ( self : List[Any] ): """simple docstring""" self.check_over_configs(variance_type=__lowercase ) self.check_over_configs(variance_type="learned_range" ) def snake_case__ ( self : List[Any] ): """simple docstring""" for num_inference_steps in [1, 2, 3, 5, 10, 50, 1_00, 9_99, 10_00]: self.check_over_forward(num_inference_steps=__lowercase , time_step=0 ) def snake_case__ ( self : int ): """simple docstring""" snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.2791 ) < 1E-3 def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = self.full_loop(use_karras_sigmas=__lowercase ) snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.2248 ) < 1E-3 def snake_case__ ( self : str ): """simple docstring""" snake_case_ = self.full_loop(prediction_type="v_prediction" ) snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.1453 ) < 1E-3 def snake_case__ ( self : int ): """simple docstring""" snake_case_ = self.full_loop(prediction_type="v_prediction" , use_karras_sigmas=__lowercase ) snake_case_ = torch.mean(torch.abs(__lowercase ) ) assert abs(result_mean.item() - 0.0649 ) < 1E-3 def snake_case__ ( self : Optional[Any] ): """simple docstring""" snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=__lowercase , dynamic_thresholding_ratio=0 ) snake_case_ = scheduler_class(**__lowercase ) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(__lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case_ = model(__lowercase , __lowercase ) snake_case_ = scheduler.step(__lowercase , __lowercase , __lowercase ).prev_sample assert sample.dtype == torch.floataa
376
1
'''simple docstring''' import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration lowerCamelCase = pytest.mark.integration lowerCamelCase = {'comet'} lowerCamelCase = importlib.util.find_spec("""fairseq""") is not None lowerCamelCase = {'code_eval'} lowerCamelCase = os.name == 'nt' lowerCamelCase = {'bertscore', 'frugalscore', 'perplexity'} lowerCamelCase = importlib.util.find_spec("""transformers""") is not None def _A ( _lowerCAmelCase ): """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowerCAmelCase ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest('"test requires Fairseq"' ) else: test_case(self , _lowercase ) return wrapper def _A ( _lowerCAmelCase ): """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowerCAmelCase ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest('"test requires transformers"' ) else: test_case(self , _lowercase ) return wrapper def _A ( _lowerCAmelCase ): """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowerCAmelCase ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest('"test not supported on Windows"' ) else: test_case(self , _lowercase ) return wrapper def _A ( ): """simple docstring""" __lowercase =[metric_dir.split(os.sep )[-2] for metric_dir in glob.glob('./metrics/*/' )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) @local class _UpperCamelCase ( parameterized.TestCase ): '''simple docstring''' lowerCAmelCase__ = {} lowerCAmelCase__ = None @pytest.mark.filterwarnings('ignore:metric_module_factory is deprecated:FutureWarning') @pytest.mark.filterwarnings('ignore:load_metric is deprecated:FutureWarning') def __lowerCamelCase ( self : Dict , _lowerCAmelCase : int): '''simple docstring''' __lowercase ='''[...]''' __lowercase =importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics' , UpperCamelCase__)).module_path) __lowercase =datasets.load.import_main_class(metric_module.__name__ , dataset=UpperCamelCase__) # check parameters __lowercase =inspect.signature(metric._compute).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values())) # no **kwargs # run doctest with self.patch_intensive_calls(UpperCamelCase__ , metric_module.__name__): with self.use_local_metrics(): try: __lowercase =doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0) self.assertGreater(results.attempted , 1) @slow def __lowerCamelCase ( self : Optional[Any] , _lowerCAmelCase : Dict): '''simple docstring''' __lowercase ='''[...]''' __lowercase =importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics' , UpperCamelCase__)).module_path) # run doctest with self.use_local_metrics(): __lowercase =doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__) self.assertEqual(results.failed , 0) self.assertGreater(results.attempted , 1) @contextmanager def __lowerCamelCase ( self : int , _lowerCAmelCase : Any , _lowerCAmelCase : List[str]): '''simple docstring''' if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](UpperCamelCase__): yield else: yield @contextmanager def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' def load_local_metric(_lowerCAmelCase : List[Any] , *_lowerCAmelCase : str , **_lowerCAmelCase : Any): return load_metric(os.path.join('metrics' , UpperCamelCase__) , *UpperCamelCase__ , **UpperCamelCase__) with patch('datasets.load_metric') as mock_load_metric: __lowercase =load_local_metric yield @classmethod def __lowerCamelCase ( cls : str , _lowerCAmelCase : Dict): '''simple docstring''' def wrapper(_lowerCAmelCase : Dict): __lowercase =contextmanager(UpperCamelCase__) __lowercase =patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher('bleurt' ) def _A ( _lowerCAmelCase ): """simple docstring""" import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string('sv' , '' , '' ) # handle pytest cli flags class _UpperCamelCase ( UpperCamelCase_ ): '''simple docstring''' def __lowerCamelCase ( self : Optional[int] , _lowerCAmelCase : str): '''simple docstring''' assert len(input_dict['input_ids']) == 2 return np.array([1.03, 1.04]) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch('bleurt.score._create_predictor' ) as mock_create_predictor: __lowercase =MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher('bertscore' ) def _A ( _lowerCAmelCase ): """simple docstring""" import torch def bert_cos_score_idf(_lowerCAmelCase , _lowerCAmelCase , *_lowerCAmelCase , **_lowerCAmelCase ): return torch.tensor([[1.0, 1.0, 1.0]] * len(_lowercase ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch('bert_score.scorer.get_model' ), patch( 'bert_score.scorer.bert_cos_score_idf' ) as mock_bert_cos_score_idf: __lowercase =bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher('comet' ) def _A ( _lowerCAmelCase ): """simple docstring""" def load_from_checkpoint(_lowerCAmelCase ): class _UpperCamelCase : '''simple docstring''' def __lowerCamelCase ( self : Any , _lowerCAmelCase : Optional[int] , *_lowerCAmelCase : List[str] , **_lowerCAmelCase : int): '''simple docstring''' assert len(UpperCamelCase__) == 2 __lowercase =[0.19, 0.92] return scores, sum(UpperCamelCase__) / len(UpperCamelCase__) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch('comet.download_model' ) as mock_download_model: __lowercase =None with patch('comet.load_from_checkpoint' ) as mock_load_from_checkpoint: __lowercase =load_from_checkpoint yield def _A ( ): """simple docstring""" __lowercase =load_metric(os.path.join('metrics' , 'seqeval' ) ) __lowercase ='''ERROR''' __lowercase =f"""Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}""" with pytest.raises(_lowercase , match=re.escape(_lowercase ) ): metric.compute(predictions=[] , references=[] , scheme=_lowercase )
712
'''simple docstring''' def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): raise ValueError('The length of profit and weight must be same.' ) if max_weight <= 0: raise ValueError('max_weight must greater than zero.' ) if any(p < 0 for p in profit ): raise ValueError('Profit can not be negative.' ) if any(w < 0 for w in weight ): raise ValueError('Weight can not be negative.' ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. __lowercase =[p / w for p, w in zip(_lowerCAmelCase , _lowerCAmelCase )] # Creating a copy of the list and sorting profit/weight in ascending order __lowercase =sorted(_lowerCAmelCase ) # declaring useful variables __lowercase =len(_lowerCAmelCase ) __lowercase =0 __lowercase =0 __lowercase =0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight __lowercase =sorted_profit_by_weight[length - i - 1] __lowercase =profit_by_weight.index(_lowerCAmelCase ) __lowercase =-1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( """Input profits, weights, and then max_weight (all positive ints) separated by """ """spaces.""" ) lowerCamelCase = [int(x) for x in input("""Input profits separated by spaces: """).split()] lowerCamelCase = [int(x) for x in input("""Input weights separated by spaces: """).split()] lowerCamelCase = int(input("""Max weight allowed: """)) # Function Call calc_profit(profit, weight, max_weight)
454
0
def lowerCAmelCase_ (lowercase__ : int = 60_08_51_47_51_43 ) -> int: '''simple docstring''' try: lowerCAmelCase__ = int(lowercase__ ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) lowerCAmelCase__ = 1 lowerCAmelCase__ = 2 while i * i <= n: while n % i == 0: lowerCAmelCase__ = i n //= i i += 1 if n > 1: lowerCAmelCase__ = n return int(lowercase__ ) if __name__ == "__main__": print(F'''{solution() = }''')
668
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _a ( a :Dict[str, torch.Tensor] ) -> Dict[str, torch.Tensor]: a = [] a = [] a = [] for rt in rc.restypes: a = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) a = {name: i for i, name in enumerate(a )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) a = torch.tensor( a , dtype=torch.intaa , device=protein['''aatype'''].device , ) a = torch.tensor( a , dtype=torch.intaa , device=protein['''aatype'''].device , ) a = torch.tensor( a , dtype=torch.floataa , device=protein['''aatype'''].device , ) a = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein a = restype_atomaa_to_atomaa[protein_aatype] a = restype_atomaa_mask[protein_aatype] a = residx_atomaa_mask a = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back a = restype_atomaa_to_atomaa[protein_aatype] a = residx_atomaa_to_atomaa.long() # create the corresponding mask a = torch.zeros([21, 37] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): a = rc.restype_atoa[restype_letter] a = rc.residue_atoms[restype_name] for atom_name in atom_names: a = rc.atom_order[atom_name] a = 1 a = restype_atomaa_mask[protein_aatype] a = residx_atomaa_mask return protein def _a ( a :Dict[str, torch.Tensor] ) -> Dict[str, np.ndarray]: a = tree_map(lambda a : torch.tensor(a , device=batch['''aatype'''].device ) , a , np.ndarray ) a = tensor_tree_map(lambda a : np.array(a ) , make_atomaa_masks(a ) ) return out
117
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class A__ : """simple docstring""" UpperCamelCase_ : CommonSchedulerState # setable values UpperCamelCase_ : jnp.ndarray UpperCamelCase_ : jnp.ndarray UpperCamelCase_ : Optional[int] = None @classmethod def _lowerCAmelCase ( cls : Tuple , lowerCAmelCase__ : CommonSchedulerState , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : jnp.ndarray ) -> Union[str, Any]: """simple docstring""" return cls(common=lowerCAmelCase__ , init_noise_sigma=lowerCAmelCase__ , timesteps=lowerCAmelCase__ ) @dataclass class A__ ( UpperCamelCase ): """simple docstring""" UpperCamelCase_ : DDPMSchedulerState class A__ ( UpperCamelCase , UpperCamelCase ): """simple docstring""" UpperCamelCase_ : str = [e.name for e in FlaxKarrasDiffusionSchedulers] UpperCamelCase_ : jnp.dtype @property def _lowerCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" return True @register_to_config def __init__( self : Dict , lowerCAmelCase__ : int = 1_0_0_0 , lowerCAmelCase__ : float = 0.0001 , lowerCAmelCase__ : float = 0.02 , lowerCAmelCase__ : str = "linear" , lowerCAmelCase__ : Optional[jnp.ndarray] = None , lowerCAmelCase__ : str = "fixed_small" , lowerCAmelCase__ : bool = True , lowerCAmelCase__ : str = "epsilon" , lowerCAmelCase__ : jnp.dtype = jnp.floataa , ) -> Optional[int]: """simple docstring""" _UpperCAmelCase : Optional[int] = dtype def _lowerCAmelCase ( self : Optional[int] , lowerCAmelCase__ : Optional[CommonSchedulerState] = None ) -> DDPMSchedulerState: """simple docstring""" if common is None: _UpperCAmelCase : List[str] = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution _UpperCAmelCase : Optional[Any] = jnp.array(1.0 , dtype=self.dtype ) _UpperCAmelCase : Dict = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=lowerCAmelCase__ , init_noise_sigma=lowerCAmelCase__ , timesteps=lowerCAmelCase__ , ) def _lowerCAmelCase ( self : Union[str, Any] , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : Optional[int] = None ) -> jnp.ndarray: """simple docstring""" return sample def _lowerCAmelCase ( self : Union[str, Any] , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : int , lowerCAmelCase__ : Tuple = () ) -> DDPMSchedulerState: """simple docstring""" _UpperCAmelCase : Any = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 _UpperCAmelCase : Dict = (jnp.arange(0 , lowerCAmelCase__ ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=lowerCAmelCase__ , timesteps=lowerCAmelCase__ , ) def _lowerCAmelCase ( self : int , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Any=None , lowerCAmelCase__ : Optional[int]=None ) -> List[Any]: """simple docstring""" _UpperCAmelCase : Union[str, Any] = state.common.alphas_cumprod[t] _UpperCAmelCase : str = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _UpperCAmelCase : int = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: _UpperCAmelCase : int = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": _UpperCAmelCase : Any = jnp.clip(lowerCAmelCase__ , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": _UpperCAmelCase : Union[str, Any] = jnp.log(jnp.clip(lowerCAmelCase__ , a_min=1e-20 ) ) elif variance_type == "fixed_large": _UpperCAmelCase : Any = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log _UpperCAmelCase : Optional[Any] = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": _UpperCAmelCase : List[Any] = variance _UpperCAmelCase : Union[str, Any] = state.common.betas[t] _UpperCAmelCase : Optional[int] = (predicted_variance + 1) / 2 _UpperCAmelCase : Any = frac * max_log + (1 - frac) * min_log return variance def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : int , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : Optional[jax.random.KeyArray] = None , lowerCAmelCase__ : bool = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: """simple docstring""" _UpperCAmelCase : int = timestep if key is None: _UpperCAmelCase : Optional[int] = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: _UpperCAmelCase : List[Any] = jnp.split(lowerCAmelCase__ , sample.shape[1] , axis=1 ) else: _UpperCAmelCase : Optional[int] = None # 1. compute alphas, betas _UpperCAmelCase : Tuple = state.common.alphas_cumprod[t] _UpperCAmelCase : Union[str, Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) _UpperCAmelCase : Any = 1 - alpha_prod_t _UpperCAmelCase : Dict = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _UpperCAmelCase : Union[str, Any] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _UpperCAmelCase : Optional[Any] = model_output elif self.config.prediction_type == "v_prediction": _UpperCAmelCase : str = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` """ " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: _UpperCAmelCase : Any = jnp.clip(lowerCAmelCase__ , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _UpperCAmelCase : List[Any] = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t _UpperCAmelCase : Tuple = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _UpperCAmelCase : Any = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): _UpperCAmelCase : List[Any] = jax.random.split(lowerCAmelCase__ , num=1 ) _UpperCAmelCase : str = jax.random.normal(lowerCAmelCase__ , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(lowerCAmelCase__ , lowerCAmelCase__ , predicted_variance=lowerCAmelCase__ ) ** 0.5) * noise _UpperCAmelCase : List[str] = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) _UpperCAmelCase : Optional[Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=lowerCAmelCase__ , state=lowerCAmelCase__ ) def _lowerCAmelCase ( self : str , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : jnp.ndarray , ) -> jnp.ndarray: """simple docstring""" return add_noise_common(state.common , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowerCAmelCase ( self : Dict , lowerCAmelCase__ : DDPMSchedulerState , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : jnp.ndarray , lowerCAmelCase__ : jnp.ndarray , ) -> jnp.ndarray: """simple docstring""" return get_velocity_common(state.common , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __len__( self : Tuple ) -> Any: """simple docstring""" return self.config.num_train_timesteps
711
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A__ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" UpperCamelCase_ : Dict = AltDiffusionPipeline UpperCamelCase_ : int = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ : Union[str, Any] = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ : str = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCamelCase_ : int = TEXT_TO_IMAGE_IMAGE_PARAMS def _lowerCAmelCase ( self : Any ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) _UpperCAmelCase : Any = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=3_2 , ) _UpperCAmelCase : Tuple = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase__ , set_alpha_to_one=lowerCAmelCase__ , ) torch.manual_seed(0 ) _UpperCAmelCase : int = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) _UpperCAmelCase : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_0_0_2 , ) _UpperCAmelCase : Dict = CLIPTextModel(lowerCAmelCase__ ) _UpperCAmelCase : str = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) _UpperCAmelCase : List[Any] = 7_7 _UpperCAmelCase : Union[str, Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def _lowerCAmelCase ( self : List[str] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Union[str, Any]=0 ) -> Tuple: """simple docstring""" if str(lowerCAmelCase__ ).startswith("mps" ): _UpperCAmelCase : int = torch.manual_seed(lowerCAmelCase__ ) else: _UpperCAmelCase : List[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def _lowerCAmelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def _lowerCAmelCase ( self : Dict ) -> List[Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def _lowerCAmelCase ( self : Optional[Any] ) -> Tuple: """simple docstring""" _UpperCAmelCase : int = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase : Union[str, Any] = self.get_dummy_components() torch.manual_seed(0 ) _UpperCAmelCase : Tuple = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5_0_0_2 , ) # TODO: remove after fixing the non-deterministic text encoder _UpperCAmelCase : int = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) _UpperCAmelCase : List[str] = text_encoder _UpperCAmelCase : int = AltDiffusionPipeline(**lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : Any = self.get_dummy_inputs(lowerCAmelCase__ ) _UpperCAmelCase : List[Any] = "A photo of an astronaut" _UpperCAmelCase : Tuple = alt_pipe(**lowerCAmelCase__ ) _UpperCAmelCase : str = output.images _UpperCAmelCase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _UpperCAmelCase : int = np.array( [0.574_8162, 0.6044_7145, 0.4882_1217, 0.5010_0636, 0.543_1185, 0.4576_3683, 0.4965_7696, 0.4813_2733, 0.4757_3093] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowerCAmelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase : Optional[int] = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase : str = self.get_dummy_components() _UpperCAmelCase : Any = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) torch.manual_seed(0 ) _UpperCAmelCase : Any = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5_0_0_2 , ) # TODO: remove after fixing the non-deterministic text encoder _UpperCAmelCase : List[Any] = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = text_encoder _UpperCAmelCase : Any = AltDiffusionPipeline(**lowerCAmelCase__ ) _UpperCAmelCase : List[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : List[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) _UpperCAmelCase : Tuple = alt_pipe(**lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = output.images _UpperCAmelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _UpperCAmelCase : Any = np.array( [0.5160_5093, 0.570_7241, 0.4736_5507, 0.5057_8886, 0.563_3877, 0.464_2503, 0.518_2081, 0.4876_3484, 0.4908_4237] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class A__ ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCAmelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" _UpperCAmelCase : Optional[Any] = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , safety_checker=lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : List[Any] = "A painting of a squirrel eating a burger" _UpperCAmelCase : Optional[Any] = torch.manual_seed(0 ) _UpperCAmelCase : Tuple = alt_pipe([prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2_0 , output_type="np" ) _UpperCAmelCase : List[str] = output.images _UpperCAmelCase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _UpperCAmelCase : List[str] = np.array([0.1010, 0.0800, 0.0794, 0.0885, 0.0843, 0.0762, 0.0769, 0.0729, 0.0586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowerCAmelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" _UpperCAmelCase : Tuple = DDIMScheduler.from_pretrained("BAAI/AltDiffusion" , subfolder="scheduler" ) _UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , scheduler=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = "A painting of a squirrel eating a burger" _UpperCAmelCase : str = torch.manual_seed(0 ) _UpperCAmelCase : Tuple = alt_pipe([prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="numpy" ) _UpperCAmelCase : List[str] = output.images _UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _UpperCAmelCase : str = np.array([0.4019, 0.4052, 0.3810, 0.4119, 0.3916, 0.3982, 0.4651, 0.4195, 0.5323] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
257
0
"""simple docstring""" from math import pi def lowerCamelCase_ (UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Dict ): return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
506
"""simple docstring""" import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup _lowercase = logging.get_logger(__name__) class __a ( __a ): '''simple docstring''' def __init__( self , **_lowerCamelCase ) -> List[Any]: '''simple docstring''' requires_backends(self , ["bs4"] ) super().__init__(**_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( self , _lowerCamelCase ) -> Tuple: '''simple docstring''' __lowercase = [] __lowercase = [] __lowercase = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag __lowercase = parent.find_all(child.name , recursive=_lowerCamelCase ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(_lowerCamelCase ) else next(i for i, s in enumerate(_lowerCamelCase , 1 ) if s is child ) ) __lowercase = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def SCREAMING_SNAKE_CASE ( self , _lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' __lowercase = BeautifulSoup(_lowerCamelCase , "html.parser" ) __lowercase = [] __lowercase = [] __lowercase = [] for element in html_code.descendants: if type(_lowerCamelCase ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue __lowercase = html.unescape(_lowerCamelCase ).strip() if not text_in_this_tag: continue all_doc_strings.append(_lowerCamelCase ) __lowercase , __lowercase = self.xpath_soup(_lowerCamelCase ) stringaxtag_seq.append(_lowerCamelCase ) stringaxsubs_seq.append(_lowerCamelCase ) if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError("Number of doc strings and xtags does not correspond" ) if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError("Number of doc strings and xsubs does not correspond" ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def SCREAMING_SNAKE_CASE ( self , _lowerCamelCase , _lowerCamelCase ) -> Tuple: '''simple docstring''' __lowercase = "" for tagname, subs in zip(_lowerCamelCase , _lowerCamelCase ): xpath += f'''/{tagname}''' if subs != 0: xpath += f'''[{subs}]''' return xpath def __call__( self , _lowerCamelCase ) -> BatchFeature: '''simple docstring''' __lowercase = False # Check that strings has a valid type if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowercase = True elif isinstance(_lowerCamelCase , (list, tuple) ): if len(_lowerCamelCase ) == 0 or isinstance(html_strings[0] , _lowerCamelCase ): __lowercase = True if not valid_strings: raise ValueError( "HTML strings must of type `str`, `List[str]` (batch of examples), " f'''but is of type {type(_lowerCamelCase )}.''' ) __lowercase = bool(isinstance(_lowerCamelCase , (list, tuple) ) and (isinstance(html_strings[0] , _lowerCamelCase )) ) if not is_batched: __lowercase = [html_strings] # Get nodes + xpaths __lowercase = [] __lowercase = [] for html_string in html_strings: __lowercase , __lowercase , __lowercase = self.get_three_from_single(_lowerCamelCase ) nodes.append(_lowerCamelCase ) __lowercase = [] for node, tag_list, sub_list in zip(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): __lowercase = self.construct_xpath(_lowerCamelCase , _lowerCamelCase ) xpath_strings.append(_lowerCamelCase ) xpaths.append(_lowerCamelCase ) # return as Dict __lowercase = {"nodes": nodes, "xpaths": xpaths} __lowercase = BatchFeature(data=_lowerCamelCase , tensor_type=_lowerCamelCase ) return encoded_inputs
118
0
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : Union[str, Any] = logging.get_logger(__name__) snake_case : List[Any] = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json''' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class snake_case_ (lowerCamelCase_ ): UpperCAmelCase__ : str = '''trocr''' UpperCAmelCase__ : List[Any] = ['''past_key_values'''] UpperCAmelCase__ : Optional[Any] = { '''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model''', '''num_hidden_layers''': '''decoder_layers''', } def __init__( self :Any ,__snake_case :int=5_02_65 ,__snake_case :Dict=10_24 ,__snake_case :List[str]=12 ,__snake_case :List[Any]=16 ,__snake_case :Optional[int]=40_96 ,__snake_case :Union[str, Any]="gelu" ,__snake_case :Optional[int]=5_12 ,__snake_case :Tuple=0.1 ,__snake_case :Any=0.0 ,__snake_case :str=0.0 ,__snake_case :Optional[Any]=2 ,__snake_case :Union[str, Any]=0.02 ,__snake_case :List[str]=0.0 ,__snake_case :Union[str, Any]=True ,__snake_case :List[Any]=False ,__snake_case :Dict=True ,__snake_case :List[Any]=True ,__snake_case :Union[str, Any]=1 ,__snake_case :List[str]=0 ,__snake_case :List[Any]=2 ,**__snake_case :List[Any] ,) -> Optional[Any]: a__ = vocab_size a__ = d_model a__ = decoder_layers a__ = decoder_attention_heads a__ = decoder_ffn_dim a__ = activation_function a__ = max_position_embeddings a__ = dropout a__ = attention_dropout a__ = activation_dropout a__ = init_std a__ = decoder_layerdrop a__ = use_cache a__ = scale_embedding a__ = use_learned_position_embeddings a__ = layernorm_embedding super().__init__( pad_token_id=__snake_case ,bos_token_id=__snake_case ,eos_token_id=__snake_case ,decoder_start_token_id=__snake_case ,**__snake_case ,)
657
from math import pi def __lowercase ( __lowerCAmelCase : int , __lowerCAmelCase : int ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
657
1
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _snake_case ( _A ): _A = 42 _A = 42 def __init__( self ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: super().__init__() self.register_modules(unet=UpperCamelCase ,scheduler=UpperCamelCase ) @torch.no_grad() def __call__( self ,UpperCamelCase = 1 ,UpperCamelCase = 2_000 ,UpperCamelCase = None ,UpperCamelCase = "pil" ,UpperCamelCase = True ,**UpperCamelCase ,) -> Union[ImagePipelineOutput, Tuple]: snake_case__ :int = self.unet.config.sample_size snake_case__ :Tuple = (batch_size, 3, img_size, img_size) snake_case__ :str = self.unet snake_case__ :int = randn_tensor(UpperCamelCase ,generator=UpperCamelCase ) * self.scheduler.init_noise_sigma snake_case__ :str = sample.to(self.device ) self.scheduler.set_timesteps(UpperCamelCase ) self.scheduler.set_sigmas(UpperCamelCase ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): snake_case__ :Union[str, Any] = self.scheduler.sigmas[i] * torch.ones(shape[0] ,device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): snake_case__ :Dict = self.unet(UpperCamelCase ,UpperCamelCase ).sample snake_case__ :Optional[int] = self.scheduler.step_correct(UpperCamelCase ,UpperCamelCase ,generator=UpperCamelCase ).prev_sample # prediction step snake_case__ :str = model(UpperCamelCase ,UpperCamelCase ).sample snake_case__ :int = self.scheduler.step_pred(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,generator=UpperCamelCase ) snake_case__ , snake_case__ :Union[str, Any] = output.prev_sample, output.prev_sample_mean snake_case__ :Dict = sample_mean.clamp(0 ,1 ) snake_case__ :Any = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": snake_case__ :Union[str, Any] = self.numpy_to_pil(UpperCamelCase ) if not return_dict: return (sample,) return ImagePipelineOutput(images=UpperCamelCase )
241
import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEncoder, BertModel, BertPreTrainedModel, ) __UpperCAmelCase : Optional[int] = logging.getLogger(__name__) class _snake_case ( _A ): def lowerCAmelCase_ ( self ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=None ,UpperCamelCase=None ) -> Any: snake_case__ :Any = self.layer[current_layer](UpperCamelCase ,UpperCamelCase ,head_mask[current_layer] ) snake_case__ :int = layer_outputs[0] return hidden_states @add_start_docstrings( 'The bare Bert Model transformer with PABEE outputting raw hidden-states without any specific head on top.' , _A , ) class _snake_case ( _A ): def __init__( self ,UpperCamelCase ) -> Any: super().__init__(UpperCamelCase ) snake_case__ :Tuple = BertEncoderWithPabee(UpperCamelCase ) self.init_weights() snake_case__ :Tuple = 0 snake_case__ :Union[str, Any] = 0 snake_case__ :Tuple = 0 snake_case__ :Union[str, Any] = 0 def lowerCAmelCase_ ( self ,UpperCamelCase ) -> Dict: snake_case__ :Tuple = threshold def lowerCAmelCase_ ( self ,UpperCamelCase ) -> Any: snake_case__ :Union[str, Any] = patience def lowerCAmelCase_ ( self ) -> List[Any]: snake_case__ :Optional[int] = 0 snake_case__ :List[Any] = 0 def lowerCAmelCase_ ( self ) -> Union[str, Any]: snake_case__ :Optional[Any] = self.inference_layers_num / self.inference_instances_num snake_case__ :Optional[Any] = ( f'*** Patience = {self.patience} Avg. Inference Layers = {avg_inf_layers:.2f} Speed Up =' f' {1 - avg_inf_layers / self.config.num_hidden_layers:.2f} ***' ) print(UpperCamelCase ) @add_start_docstrings_to_model_forward(UpperCamelCase ) def lowerCAmelCase_ ( self ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=False ,) -> str: if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time" ) elif input_ids is not None: snake_case__ :List[str] = input_ids.size() elif inputs_embeds is not None: snake_case__ :List[str] = inputs_embeds.size()[:-1] else: raise ValueError("You have to specify either input_ids or inputs_embeds" ) snake_case__ :List[Any] = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: snake_case__ :Union[str, Any] = torch.ones(UpperCamelCase ,device=UpperCamelCase ) if token_type_ids is None: snake_case__ :Any = torch.zeros(UpperCamelCase ,dtype=torch.long ,device=UpperCamelCase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. snake_case__ :torch.Tensor = self.get_extended_attention_mask(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if self.config.is_decoder and encoder_hidden_states is not None: snake_case__ , snake_case__ , snake_case__ :Union[str, Any] = encoder_hidden_states.size() snake_case__ :int = (encoder_batch_size, encoder_sequence_length) if encoder_attention_mask is None: snake_case__ :Optional[int] = torch.ones(UpperCamelCase ,device=UpperCamelCase ) snake_case__ :Optional[int] = self.invert_attention_mask(UpperCamelCase ) else: snake_case__ :Dict = None # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] snake_case__ :Tuple = self.get_head_mask(UpperCamelCase ,self.config.num_hidden_layers ) snake_case__ :List[str] = self.embeddings( input_ids=UpperCamelCase ,position_ids=UpperCamelCase ,token_type_ids=UpperCamelCase ,inputs_embeds=UpperCamelCase ) snake_case__ :Union[str, Any] = embedding_output if self.training: snake_case__ :Optional[int] = [] for i in range(self.config.num_hidden_layers ): snake_case__ :List[Any] = self.encoder.adaptive_forward( UpperCamelCase ,current_layer=UpperCamelCase ,attention_mask=UpperCamelCase ,head_mask=UpperCamelCase ) snake_case__ :int = self.pooler(UpperCamelCase ) snake_case__ :Tuple = output_layers[i](output_dropout(UpperCamelCase ) ) res.append(UpperCamelCase ) elif self.patience == 0: # Use all layers for inference snake_case__ :Any = self.encoder( UpperCamelCase ,attention_mask=UpperCamelCase ,head_mask=UpperCamelCase ,encoder_hidden_states=UpperCamelCase ,encoder_attention_mask=UpperCamelCase ,) snake_case__ :Optional[int] = self.pooler(encoder_outputs[0] ) snake_case__ :Optional[int] = [output_layers[self.config.num_hidden_layers - 1](UpperCamelCase )] else: snake_case__ :Optional[int] = 0 snake_case__ :Dict = None snake_case__ :str = 0 for i in range(self.config.num_hidden_layers ): calculated_layer_num += 1 snake_case__ :Any = self.encoder.adaptive_forward( UpperCamelCase ,current_layer=UpperCamelCase ,attention_mask=UpperCamelCase ,head_mask=UpperCamelCase ) snake_case__ :Union[str, Any] = self.pooler(UpperCamelCase ) snake_case__ :Optional[Any] = output_layers[i](UpperCamelCase ) if regression: snake_case__ :Optional[Any] = logits.detach() if patient_result is not None: snake_case__ :Union[str, Any] = patient_result.detach() if (patient_result is not None) and torch.abs(patient_result - labels ) < self.regression_threshold: patient_counter += 1 else: snake_case__ :Optional[Any] = 0 else: snake_case__ :Tuple = logits.detach().argmax(dim=1 ) if patient_result is not None: snake_case__ :Optional[int] = patient_result.detach().argmax(dim=1 ) if (patient_result is not None) and torch.all(labels.eq(UpperCamelCase ) ): patient_counter += 1 else: snake_case__ :Any = 0 snake_case__ :int = logits if patient_counter == self.patience: break snake_case__ :int = [patient_result] self.inference_layers_num += calculated_layer_num self.inference_instances_num += 1 return res @add_start_docstrings( 'Bert Model transformer with PABEE and a sequence classification/regression head on top (a linear layer on top of\n the pooled output) e.g. for GLUE tasks. ' , _A , ) class _snake_case ( _A ): def __init__( self ,UpperCamelCase ) -> Union[str, Any]: super().__init__(UpperCamelCase ) snake_case__ :Optional[int] = config.num_labels snake_case__ :Tuple = BertModelWithPabee(UpperCamelCase ) snake_case__ :Any = nn.Dropout(config.hidden_dropout_prob ) snake_case__ :Optional[int] = nn.ModuleList( [nn.Linear(config.hidden_size ,self.config.num_labels ) for _ in range(config.num_hidden_layers )] ) self.init_weights() @add_start_docstrings_to_model_forward(UpperCamelCase ) def lowerCAmelCase_ ( self ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase=None ,) -> Any: snake_case__ :Optional[int] = self.bert( input_ids=UpperCamelCase ,attention_mask=UpperCamelCase ,token_type_ids=UpperCamelCase ,position_ids=UpperCamelCase ,head_mask=UpperCamelCase ,inputs_embeds=UpperCamelCase ,output_dropout=self.dropout ,output_layers=self.classifiers ,regression=self.num_labels == 1 ,) snake_case__ :Tuple = (logits[-1],) if labels is not None: snake_case__ :Optional[int] = None snake_case__ :Optional[Any] = 0 for ix, logits_item in enumerate(UpperCamelCase ): if self.num_labels == 1: # We are doing regression snake_case__ :int = MSELoss() snake_case__ :str = loss_fct(logits_item.view(-1 ) ,labels.view(-1 ) ) else: snake_case__ :int = CrossEntropyLoss() snake_case__ :Union[str, Any] = loss_fct(logits_item.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) if total_loss is None: snake_case__ :Optional[Any] = loss else: total_loss += loss * (ix + 1) total_weights += ix + 1 snake_case__ :List[Any] = (total_loss / total_weights,) + outputs return outputs
241
1
from ...processing_utils import ProcessorMixin class _a ( _a ): '''simple docstring''' lowerCamelCase_ : Any = ["""image_processor""", """feature_extractor"""] lowerCamelCase_ : int = """TvltImageProcessor""" lowerCamelCase_ : Dict = """TvltFeatureExtractor""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): super().__init__(image_processor=_A , feature_extractor=_A ) __A : Dict = image_processor __A : str = feature_extractor def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=False , __UpperCAmelCase=False , *__UpperCAmelCase , **__UpperCAmelCase , ): if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) __A : Union[str, Any] = None if images is not None: __A : Tuple = self.image_processor(_A , mask_pixel=_A , *_A , **_A ) if images_mixed is not None: __A : Union[str, Any] = self.image_processor(_A , is_mixed=_A , *_A , **_A ) if audio is not None: __A : Optional[int] = self.feature_extractor( _A , *_A , sampling_rate=_A , mask_audio=_A , **_A ) __A : List[Any] = {} if audio is not None: output_dict.update(_A ) if images is not None: output_dict.update(_A ) if images_mixed_dict is not None: output_dict.update(_A ) return output_dict @property def __UpperCAmelCase( self ): __A : Optional[Any] = self.image_processor.model_input_names __A : List[str] = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
714
import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class _a ( lowerCAmelCase__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): super().__init__() if safety_checker is None: logger.warning( F"You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure" " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( speech_model=__UpperCAmelCase , speech_processor=__UpperCAmelCase , vae=__UpperCAmelCase , text_encoder=__UpperCAmelCase , tokenizer=__UpperCAmelCase , unet=__UpperCAmelCase , scheduler=__UpperCAmelCase , feature_extractor=__UpperCAmelCase , ) def __UpperCAmelCase( self , __UpperCAmelCase = "auto" ): if slice_size == "auto": __A : int = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__UpperCAmelCase ) def __UpperCAmelCase( self ): self.enable_attention_slicing(__UpperCAmelCase ) @torch.no_grad() def __call__( self , __UpperCAmelCase , __UpperCAmelCase=16_000 , __UpperCAmelCase = 512 , __UpperCAmelCase = 512 , __UpperCAmelCase = 50 , __UpperCAmelCase = 7.5 , __UpperCAmelCase = None , __UpperCAmelCase = 1 , __UpperCAmelCase = 0.0 , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = "pil" , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = 1 , **__UpperCAmelCase , ): __A : List[str] = self.speech_processor.feature_extractor( __UpperCAmelCase , return_tensors="pt" , sampling_rate=__UpperCAmelCase ).input_features.to(self.device ) __A : Any = self.speech_model.generate(__UpperCAmelCase , max_length=480_000 ) __A : List[str] = self.speech_processor.tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase , normalize=__UpperCAmelCase )[ 0 ] if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __A : Optional[Any] = 1 elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): __A : Dict = len(__UpperCAmelCase ) else: raise ValueError(F"`prompt` has to be of type `str` or `list` but is {type(__UpperCAmelCase )}" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or callback_steps <= 0) ): raise ValueError( F"`callback_steps` has to be a positive integer but is {callback_steps} of type" F" {type(__UpperCAmelCase )}." ) # get prompt text embeddings __A : Optional[int] = self.tokenizer( __UpperCAmelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , return_tensors="pt" , ) __A : int = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: __A : List[str] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" F" {self.tokenizer.model_max_length} tokens: {removed_text}" ) __A : Dict = text_input_ids[:, : self.tokenizer.model_max_length] __A : int = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method __A , __A , __A : str = text_embeddings.shape __A : Optional[int] = text_embeddings.repeat(1 , __UpperCAmelCase , 1 ) __A : List[str] = text_embeddings.view(bs_embed * num_images_per_prompt , __UpperCAmelCase , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __A : Dict = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __A : List[str] if negative_prompt is None: __A : Dict = [""] * batch_size elif type(__UpperCAmelCase ) is not type(__UpperCAmelCase ): raise TypeError( F"`negative_prompt` should be the same type to `prompt`, but got {type(__UpperCAmelCase )} !=" F" {type(__UpperCAmelCase )}." ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): __A : Any = [negative_prompt] elif batch_size != len(__UpperCAmelCase ): raise ValueError( F"`negative_prompt`: {negative_prompt} has batch size {len(__UpperCAmelCase )}, but `prompt`:" F" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches" " the batch size of `prompt`." ) else: __A : int = negative_prompt __A : int = text_input_ids.shape[-1] __A : Any = self.tokenizer( __UpperCAmelCase , padding="max_length" , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , return_tensors="pt" , ) __A : int = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __A : Union[str, Any] = uncond_embeddings.shape[1] __A : List[str] = uncond_embeddings.repeat(1 , __UpperCAmelCase , 1 ) __A : int = uncond_embeddings.view(batch_size * num_images_per_prompt , __UpperCAmelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __A : Optional[int] = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __A : Dict = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) __A : Any = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps __A : Tuple = torch.randn(__UpperCAmelCase , generator=__UpperCAmelCase , device="cpu" , dtype=__UpperCAmelCase ).to( self.device ) else: __A : List[Any] = torch.randn(__UpperCAmelCase , generator=__UpperCAmelCase , device=self.device , dtype=__UpperCAmelCase ) else: if latents.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) __A : Tuple = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCAmelCase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand __A : Optional[int] = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __A : Tuple = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __A : Any = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __A : List[str] = {} if accepts_eta: __A : Tuple = eta for i, t in enumerate(self.progress_bar(__UpperCAmelCase ) ): # expand the latents if we are doing classifier free guidance __A : Union[str, Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = self.scheduler.scale_model_input(__UpperCAmelCase , __UpperCAmelCase ) # predict the noise residual __A : List[Any] = self.unet(__UpperCAmelCase , __UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase ).sample # perform guidance if do_classifier_free_guidance: __A , __A : str = noise_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 __A : Union[str, Any] = self.scheduler.step(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) __A : int = 1 / 0.1_82_15 * latents __A : Union[str, Any] = self.vae.decode(__UpperCAmelCase ).sample __A : Tuple = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 __A : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : List[str] = self.numpy_to_pil(__UpperCAmelCase ) if not return_dict: return image return StableDiffusionPipelineOutput(images=__UpperCAmelCase , nsfw_content_detected=__UpperCAmelCase )
387
0
"""simple docstring""" import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowercase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" def __init__( self , _A , _A , _A ): '''simple docstring''' super().__init__() self.register_modules(vqvae=_A , unet=_A , scheduler=_A ) @torch.no_grad() def __call__( self , _A = 1 , _A = None , _A = 0.0 , _A = 5_0 , _A = "pil" , _A = True , **_A , ): '''simple docstring''' UpperCamelCase : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=_A , ) UpperCamelCase : Optional[int] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler UpperCamelCase : Union[str, Any] = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_A ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature UpperCamelCase : Dict = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCamelCase : int = {} if accepts_eta: UpperCamelCase : Any = eta for t in self.progress_bar(self.scheduler.timesteps ): UpperCamelCase : Tuple = self.scheduler.scale_model_input(_A , _A ) # predict the noise residual UpperCamelCase : str = self.unet(_A , _A ).sample # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase : Tuple = self.scheduler.step(_A , _A , _A , **_A ).prev_sample # decode the image latents with the VAE UpperCamelCase : Tuple = self.vqvae.decode(_A ).sample UpperCamelCase : Optional[Any] = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase : Optional[int] = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
102
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowercase__ : """simple docstring""" def __init__( self , _A , _A=1_3 , _A=2 , _A=2_4 , _A=1_6 , _A=True , _A=True , _A=3_2 , _A=5 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=1_0 , _A=0.02 , _A=None , _A=2 , _A=2 , ): '''simple docstring''' UpperCamelCase : Tuple = parent UpperCamelCase : str = batch_size UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : Optional[int] = max_length UpperCamelCase : str = num_mel_bins UpperCamelCase : Optional[int] = is_training UpperCamelCase : str = use_labels UpperCamelCase : Optional[int] = hidden_size UpperCamelCase : Union[str, Any] = num_hidden_layers UpperCamelCase : Optional[Any] = num_attention_heads UpperCamelCase : Union[str, Any] = intermediate_size UpperCamelCase : str = hidden_act UpperCamelCase : Optional[Any] = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = type_sequence_label_size UpperCamelCase : List[Any] = initializer_range UpperCamelCase : Optional[Any] = scope UpperCamelCase : List[Any] = frequency_stride UpperCamelCase : Tuple = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) UpperCamelCase : Optional[Any] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 UpperCamelCase : Dict = (self.max_length - self.patch_size) // self.time_stride + 1 UpperCamelCase : List[str] = frequency_out_dimension * time_out_dimension UpperCamelCase : int = num_patches + 2 def _a ( self ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) UpperCamelCase : int = None if self.use_labels: UpperCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Dict = self.get_config() return config, input_values, labels def _a ( self ): '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_A , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def _a ( self , _A , _A , _A ): '''simple docstring''' UpperCamelCase : List[str] = ASTModel(config=_A ) model.to(_A ) model.eval() UpperCamelCase : int = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : List[str] = config_and_inputs UpperCamelCase : Union[str, Any] = {"""input_values""": input_values} return config, inputs_dict @require_torch class lowercase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" __lowerCAmelCase : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) __lowerCAmelCase : Any = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) __lowerCAmelCase : Union[str, Any] = False __lowerCAmelCase : Any = False __lowerCAmelCase : List[Any] = False __lowerCAmelCase : Union[str, Any] = False def _a ( self , _A , _A , _A , _A , _A ): '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def _a ( self ): '''simple docstring''' UpperCamelCase : str = ASTModelTester(self ) UpperCamelCase : int = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=3_7 ) def _a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""AST does not use inputs_embeds""" ) def _a ( self ): '''simple docstring''' pass def _a ( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Tuple = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear ) ) def _a ( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Dict = model_class(_A ) UpperCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : int = [*signature.parameters.keys()] UpperCamelCase : Any = ["""input_values"""] self.assertListEqual(arg_names[:1] , _A ) def _a ( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) @slow def _a ( self ): '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : int = ASTModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def UpperCamelCase (): UpperCamelCase : Optional[Any] = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" ) UpperCamelCase , UpperCamelCase : Dict = torchaudio.load(SCREAMING_SNAKE_CASE ) return audio, sampling_rate @require_torch @require_torchaudio class lowercase__ ( unittest.TestCase ): """simple docstring""" @cached_property def _a ( self ): '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ) if is_torchaudio_available() else None ) @slow def _a ( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.default_feature_extractor UpperCamelCase : Any = ASTForAudioClassification.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ).to(_A ) UpperCamelCase : Union[str, Any] = self.default_feature_extractor UpperCamelCase , UpperCamelCase : List[Any] = prepare_audio() UpperCamelCase : Optional[Any] = audio.squeeze().numpy() UpperCamelCase : Union[str, Any] = feature_extractor(_A , sampling_rate=_A , return_tensors="""pt""" ).to(_A ) # forward pass with torch.no_grad(): UpperCamelCase : Optional[Any] = model(**_A ) # verify the logits UpperCamelCase : List[Any] = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , _A ) UpperCamelCase : Optional[Any] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1e-4 ) )
102
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class UpperCamelCase ( lowercase , unittest.TestCase ): UpperCAmelCase : Any = ShapEPipeline UpperCAmelCase : Union[str, Any] = ["""prompt"""] UpperCAmelCase : str = ["""prompt"""] UpperCAmelCase : List[str] = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] UpperCAmelCase : Dict = False @property def _lowercase (self : Tuple) -> str: return 32 @property def _lowercase (self : Tuple) -> Optional[int]: return 32 @property def _lowercase (self : Any) -> str: return self.time_input_dim * 4 @property def _lowercase (self : Any) -> Any: return 8 @property def _lowercase (self : List[str]) -> Dict: __snake_case : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') return tokenizer @property def _lowercase (self : Union[str, Any]) -> List[Any]: torch.manual_seed(0) __snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(_A) @property def _lowercase (self : int) -> Dict: torch.manual_seed(0) __snake_case : str = { 'num_attention_heads': 2, 'attention_head_dim': 16, 'embedding_dim': self.time_input_dim, 'num_embeddings': 32, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } __snake_case : str = PriorTransformer(**_A) return model @property def _lowercase (self : Optional[Any]) -> Union[str, Any]: torch.manual_seed(0) __snake_case : Any = { 'param_shapes': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 12, 'background': ( 0.1, 0.1, 0.1, ), } __snake_case : int = ShapERenderer(**_A) return model def _lowercase (self : List[str]) -> Optional[Any]: __snake_case : Union[str, Any] = self.dummy_prior __snake_case : Tuple = self.dummy_text_encoder __snake_case : Optional[int] = self.dummy_tokenizer __snake_case : Optional[int] = self.dummy_renderer __snake_case : Optional[int] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=10_24 , prediction_type='sample' , use_karras_sigmas=_A , clip_sample=_A , clip_sample_range=1.0 , ) __snake_case : Optional[int] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def _lowercase (self : Union[str, Any] , _A : List[str] , _A : Dict=0) -> Union[str, Any]: if str(_A).startswith('mps'): __snake_case : List[Any] = torch.manual_seed(_A) else: __snake_case : Any = torch.Generator(device=_A).manual_seed(_A) __snake_case : List[str] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 32, 'output_type': 'np', } return inputs def _lowercase (self : Union[str, Any]) -> str: __snake_case : List[Any] = 'cpu' __snake_case : Optional[int] = self.get_dummy_components() __snake_case : List[Any] = self.pipeline_class(**_A) __snake_case : Tuple = pipe.to(_A) pipe.set_progress_bar_config(disable=_A) __snake_case : Any = pipe(**self.get_dummy_inputs(_A)) __snake_case : List[str] = output.images[0] __snake_case : Any = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __snake_case : Dict = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _lowercase (self : Tuple) -> int: # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2]) def _lowercase (self : int) -> Tuple: __snake_case : List[str] = torch_device == 'cpu' __snake_case : Optional[int] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_A , relax_max_difference=_A , ) def _lowercase (self : Optional[int]) -> str: __snake_case : Dict = self.get_dummy_components() __snake_case : List[Any] = self.pipeline_class(**_A) __snake_case : Optional[int] = pipe.to(_A) pipe.set_progress_bar_config(disable=_A) __snake_case : Union[str, Any] = 1 __snake_case : List[Any] = 2 __snake_case : Dict = self.get_dummy_inputs(_A) for key in inputs.keys(): if key in self.batch_params: __snake_case : Dict = batch_size * [inputs[key]] __snake_case : str = pipe(**_A , num_images_per_prompt=_A)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : int) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase (self : List[str]) -> Union[str, Any]: __snake_case : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy') __snake_case : List[str] = ShapEPipeline.from_pretrained('openai/shap-e') __snake_case : Any = pipe.to(_A) pipe.set_progress_bar_config(disable=_A) __snake_case : Optional[int] = torch.Generator(device=_A).manual_seed(0) __snake_case : List[Any] = pipe( 'a shark' , generator=_A , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type='np' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_A , _A)
192
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType _a : Optional[int]= logging.get_logger(__name__) _a : Tuple= { "microsoft/layoutlmv3-base": "https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json", } class UpperCamelCase ( lowercase ): UpperCAmelCase : Dict = """layoutlmv3""" def __init__(self : List[str] , _A : Optional[int]=5_02_65 , _A : List[str]=7_68 , _A : List[Any]=12 , _A : List[str]=12 , _A : Optional[int]=30_72 , _A : str="gelu" , _A : int=0.1 , _A : Tuple=0.1 , _A : List[Any]=5_12 , _A : List[str]=2 , _A : List[Any]=0.02 , _A : Tuple=1E-5 , _A : Dict=1 , _A : str=0 , _A : str=2 , _A : List[str]=10_24 , _A : Optional[Any]=1_28 , _A : Union[str, Any]=1_28 , _A : Union[str, Any]=True , _A : Union[str, Any]=32 , _A : Any=1_28 , _A : Optional[Any]=64 , _A : List[Any]=2_56 , _A : str=True , _A : List[Any]=True , _A : Tuple=True , _A : Tuple=2_24 , _A : Tuple=3 , _A : Optional[Any]=16 , _A : Tuple=None , **_A : Optional[int] , ) -> Optional[int]: super().__init__( vocab_size=_A , hidden_size=_A , num_hidden_layers=_A , num_attention_heads=_A , intermediate_size=_A , hidden_act=_A , hidden_dropout_prob=_A , attention_probs_dropout_prob=_A , max_position_embeddings=_A , type_vocab_size=_A , initializer_range=_A , layer_norm_eps=_A , pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A , ) __snake_case : List[Any] = max_ad_position_embeddings __snake_case : List[str] = coordinate_size __snake_case : int = shape_size __snake_case : List[str] = has_relative_attention_bias __snake_case : Union[str, Any] = rel_pos_bins __snake_case : Tuple = max_rel_pos __snake_case : Optional[Any] = has_spatial_attention_bias __snake_case : Union[str, Any] = rel_ad_pos_bins __snake_case : Any = max_rel_ad_pos __snake_case : Tuple = text_embed __snake_case : str = visual_embed __snake_case : List[str] = input_size __snake_case : List[Any] = num_channels __snake_case : Union[str, Any] = patch_size __snake_case : str = classifier_dropout class UpperCamelCase ( lowercase ): UpperCAmelCase : Optional[int] = version.parse("""1.12""" ) @property def _lowercase (self : Optional[int]) -> Mapping[str, Mapping[int, str]]: # The order of inputs is different for question answering and sequence classification if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ]) else: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels'}), ]) @property def _lowercase (self : Tuple) -> float: return 1E-5 @property def _lowercase (self : Optional[int]) -> int: return 12 def _lowercase (self : str , _A : "ProcessorMixin" , _A : int = -1 , _A : int = -1 , _A : bool = False , _A : Optional["TensorType"] = None , _A : int = 3 , _A : int = 40 , _A : int = 40 , ) -> Mapping[str, Any]: setattr(processor.image_processor , 'apply_ocr' , _A) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __snake_case : Union[str, Any] = compute_effective_axis_dimension( _A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __snake_case : Any = processor.tokenizer.num_special_tokens_to_add(_A) __snake_case : Optional[int] = compute_effective_axis_dimension( _A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_A) # Generate dummy inputs according to compute batch and sequence __snake_case : Optional[int] = [[' '.join([processor.tokenizer.unk_token]) * seq_length]] * batch_size # Generate dummy bounding boxes __snake_case : Dict = [[[48, 84, 73, 1_28]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) __snake_case : Optional[Any] = self._generate_dummy_images(_A , _A , _A , _A) __snake_case : Dict = dict( processor( _A , text=_A , boxes=_A , return_tensors=_A , )) return inputs
192
1
from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class _A ( _lowerCamelCase ): _UpperCamelCase : Optional[int] = DistilBertTokenizer _UpperCamelCase : Optional[Any] = DistilBertTokenizerFast _UpperCamelCase : int = True @slow def __a ( self : Dict ) -> Dict: """simple docstring""" lowercase : Tuple = DistilBertTokenizer.from_pretrained('''distilbert-base-uncased''' ) lowercase : Union[str, Any] = tokenizer.encode('''sequence builders''' , add_special_tokens=_A ) lowercase : List[Any] = tokenizer.encode('''multi-sequence build''' , add_special_tokens=_A ) lowercase : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(_A ) lowercase : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_A , _A ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
217
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCAmelCase_ = TypeVar('T') class _A ( Generic[T] ): def __init__( self : Optional[Any] , _A : T ) -> Union[str, Any]: """simple docstring""" lowercase : Optional[Any] = data lowercase : Node[T] | None = None def __str__( self : List[str] ) -> str: """simple docstring""" return f"""{self.data}""" class _A ( Generic[T] ): def __init__( self : str ) -> None: """simple docstring""" lowercase : Node[T] | None = None def __iter__( self : List[str] ) -> Iterator[T]: """simple docstring""" lowercase : int = self.top while node: yield node.data lowercase : List[str] = node.next def __str__( self : Tuple ) -> str: """simple docstring""" return "->".join([str(_A ) for item in self] ) def __len__( self : Tuple ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def __a ( self : List[str] ) -> bool: """simple docstring""" return self.top is None def __a ( self : List[Any] , _A : T ) -> None: """simple docstring""" lowercase : Any = Node(_A ) if not self.is_empty(): lowercase : str = self.top lowercase : Any = node def __a ( self : List[Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError('''pop from empty stack''' ) assert isinstance(self.top , _A ) lowercase : int = self.top lowercase : Optional[int] = self.top.next return pop_node.data def __a ( self : Optional[Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError('''peek from empty stack''' ) assert self.top is not None return self.top.data def __a ( self : Dict ) -> None: """simple docstring""" lowercase : str = None if __name__ == "__main__": from doctest import testmod testmod()
217
1
from __future__ import annotations __magic_name__ : List[Any] = 8.9_8_8e9 # units = N * m^s * C^-2 def lowerCAmelCase ( snake_case__ : float , snake_case__ : float , snake_case__ : float , snake_case__ : float )-> dict[str, float]: A_ = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if distance < 0: raise ValueError("Distance cannot be negative" ) if force == 0: A_ = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: A_ = abs(snake_case__ ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: A_ = abs(snake_case__ ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: A_ = (COULOMBS_CONSTANT * charge_product / abs(snake_case__ )) ** 0.5 return {"distance": distance} raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
608
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging __magic_name__ : Optional[Any] = logging.get_logger(__name__) __magic_name__ : Tuple = { 'google/umt5-small': 'https://huggingface.co/google/umt5-small/resolve/main/config.json', # See all umt5 models at https://huggingface.co/models?filter=umt5 } class lowerCamelCase ( __snake_case ): """simple docstring""" lowerCAmelCase_ = """umt5""" lowerCAmelCase_ = ["""past_key_values"""] def __init__( self , __UpperCamelCase=250112 , __UpperCamelCase=512 , __UpperCamelCase=64 , __UpperCamelCase=1024 , __UpperCamelCase=8 , __UpperCamelCase=None , __UpperCamelCase=6 , __UpperCamelCase=32 , __UpperCamelCase=128 , __UpperCamelCase=0.1 , __UpperCamelCase=1E-6 , __UpperCamelCase=1.0 , __UpperCamelCase="gated-gelu" , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase="T5Tokenizer" , __UpperCamelCase=True , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=0 , **__UpperCamelCase , ): super().__init__( is_encoder_decoder=__UpperCamelCase , tokenizer_class=__UpperCamelCase , tie_word_embeddings=__UpperCamelCase , pad_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , decoder_start_token_id=__UpperCamelCase , **__UpperCamelCase , ) A_ = vocab_size A_ = d_model A_ = d_kv A_ = d_ff A_ = num_layers A_ = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry A_ = num_heads A_ = relative_attention_num_buckets A_ = relative_attention_max_distance A_ = dropout_rate A_ = layer_norm_epsilon A_ = initializer_factor A_ = feed_forward_proj A_ = use_cache A_ = self.feed_forward_proj.split("-" ) A_ = act_info[-1] A_ = act_info[0] == "gated" if len(__UpperCamelCase ) > 1 and act_info[0] != "gated" or len(__UpperCamelCase ) > 2: raise ValueError( f'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "'gated-gelu' or 'relu'" ) if feed_forward_proj == "gated-gelu": A_ = "gelu_new" @property def lowercase_ ( self ): return self.d_model @property def lowercase_ ( self ): return self.num_heads @property def lowercase_ ( self ): return self.num_layers class lowerCamelCase ( __snake_case ): """simple docstring""" @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def lowercase_ ( self ): A_ = { "input_ids": {0: "batch", 1: "encoder_sequence"}, "attention_mask": {0: "batch", 1: "encoder_sequence"}, } if self.use_past: A_ = "past_encoder_sequence + sequence" A_ = {0: "batch"} A_ = {0: "batch", 1: "past_decoder_sequence + sequence"} else: A_ = {0: "batch", 1: "decoder_sequence"} A_ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__UpperCamelCase , direction="inputs" ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def lowercase_ ( self ): return 13 @property def lowercase_ ( self ): return 5E-4
608
1
from collections.abc import Callable import numpy as np def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = int(np.ceil((x_end - xa) / step_size ) ) UpperCAmelCase_ : Dict = np.zeros((n + 1,) ) UpperCAmelCase_ : str = ya UpperCAmelCase_ : Tuple = xa for k in range(_lowercase ): UpperCAmelCase_ : Optional[int] = y[k] + step_size * ode_func(_lowercase , y[k] ) UpperCAmelCase_ : Optional[Any] = y[k] + ( (step_size / 2) * (ode_func(_lowercase , y[k] ) + ode_func(x + step_size , _lowercase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
30
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ : Dict = logging.get_logger(__name__) UpperCAmelCase__ : Tuple = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class lowerCAmelCase_ (a__ ): """simple docstring""" __UpperCamelCase : Dict = '''unispeech-sat''' def __init__(self , SCREAMING_SNAKE_CASE__=32 , SCREAMING_SNAKE_CASE__=7_68 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=30_72 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.02 , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__="group" , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , SCREAMING_SNAKE_CASE__=(5, 2, 2, 2, 2, 2, 2) , SCREAMING_SNAKE_CASE__=(10, 3, 3, 3, 3, 2, 2) , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=1_28 , SCREAMING_SNAKE_CASE__=16 , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=0.05 , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=3_20 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=1_00 , SCREAMING_SNAKE_CASE__=2_56 , SCREAMING_SNAKE_CASE__=2_56 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__="mean" , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=2_56 , SCREAMING_SNAKE_CASE__=(5_12, 5_12, 5_12, 5_12, 15_00) , SCREAMING_SNAKE_CASE__=(5, 3, 3, 1, 1) , SCREAMING_SNAKE_CASE__=(1, 2, 3, 1, 1) , SCREAMING_SNAKE_CASE__=5_12 , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=1 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=5_04 , **SCREAMING_SNAKE_CASE__ , ) -> List[Any]: """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE__ , pad_token_id=SCREAMING_SNAKE_CASE__ , bos_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[Any] = hidden_size SCREAMING_SNAKE_CASE__ : Optional[Any] = feat_extract_norm SCREAMING_SNAKE_CASE__ : Tuple = feat_extract_activation SCREAMING_SNAKE_CASE__ : List[Any] = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[Any] = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = conv_bias SCREAMING_SNAKE_CASE__ : List[Any] = num_conv_pos_embeddings SCREAMING_SNAKE_CASE__ : Tuple = num_conv_pos_embedding_groups SCREAMING_SNAKE_CASE__ : str = len(self.conv_dim ) SCREAMING_SNAKE_CASE__ : str = num_hidden_layers SCREAMING_SNAKE_CASE__ : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE__ : List[str] = hidden_act SCREAMING_SNAKE_CASE__ : Optional[int] = num_attention_heads SCREAMING_SNAKE_CASE__ : List[str] = hidden_dropout SCREAMING_SNAKE_CASE__ : List[Any] = attention_dropout SCREAMING_SNAKE_CASE__ : str = activation_dropout SCREAMING_SNAKE_CASE__ : List[Any] = feat_proj_dropout SCREAMING_SNAKE_CASE__ : Dict = final_dropout SCREAMING_SNAKE_CASE__ : List[str] = layerdrop SCREAMING_SNAKE_CASE__ : Tuple = layer_norm_eps SCREAMING_SNAKE_CASE__ : Optional[int] = initializer_range SCREAMING_SNAKE_CASE__ : str = vocab_size SCREAMING_SNAKE_CASE__ : Union[str, Any] = num_clusters SCREAMING_SNAKE_CASE__ : List[Any] = do_stable_layer_norm SCREAMING_SNAKE_CASE__ : List[str] = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 SCREAMING_SNAKE_CASE__ : Optional[Any] = apply_spec_augment SCREAMING_SNAKE_CASE__ : Tuple = mask_time_prob SCREAMING_SNAKE_CASE__ : str = mask_time_length SCREAMING_SNAKE_CASE__ : Optional[int] = mask_time_min_masks SCREAMING_SNAKE_CASE__ : int = mask_feature_prob SCREAMING_SNAKE_CASE__ : str = mask_feature_length SCREAMING_SNAKE_CASE__ : Optional[Any] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations SCREAMING_SNAKE_CASE__ : int = num_codevectors_per_group SCREAMING_SNAKE_CASE__ : Optional[int] = num_codevector_groups SCREAMING_SNAKE_CASE__ : List[Any] = contrastive_logits_temperature SCREAMING_SNAKE_CASE__ : Optional[Any] = feat_quantizer_dropout SCREAMING_SNAKE_CASE__ : str = num_negatives SCREAMING_SNAKE_CASE__ : List[Any] = codevector_dim SCREAMING_SNAKE_CASE__ : Optional[Any] = proj_codevector_dim SCREAMING_SNAKE_CASE__ : List[Any] = diversity_loss_weight # ctc loss SCREAMING_SNAKE_CASE__ : int = ctc_loss_reduction SCREAMING_SNAKE_CASE__ : Optional[int] = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE__ : Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE__ : Any = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = list(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = xvector_output_dim @property def __magic_name__ (self ) -> Optional[int]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
223
0
import argparse from collections import defaultdict import yaml lowerCAmelCase : Tuple = 'docs/source/en/_toctree.yml' def A_ ( a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = defaultdict(a ) SCREAMING_SNAKE_CASE_ : int = [] SCREAMING_SNAKE_CASE_ : Optional[int] = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({'local': doc['local'], 'title': doc['title']} ) else: new_doc_list.append(a ) SCREAMING_SNAKE_CASE_ : Dict = new_doc_list SCREAMING_SNAKE_CASE_ : Union[str, Any] = [key for key, value in counts.items() if value > 1] SCREAMING_SNAKE_CASE_ : Any = [] for duplicate_key in duplicates: SCREAMING_SNAKE_CASE_ : Any = list({doc['title'] for doc in doc_list if doc['local'] == duplicate_key} ) if len(a ) > 1: raise ValueError( f"{duplicate_key} is present several times in the documentation table of content at " '`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ' 'others.' ) # Only add this once new_doc.append({'local': duplicate_key, 'title': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if 'local' not in counts or counts[doc['local']] == 1] ) SCREAMING_SNAKE_CASE_ : int = sorted(a , key=lambda a : s["title"].lower() ) # "overview" gets special treatment and is always first if len(a ) > 1: raise ValueError('{doc_list} has two \'overview\' docs which is not allowed.' ) overview_doc.extend(a ) # Sort return overview_doc def A_ ( a=False ): """simple docstring""" with open(a , encoding='utf-8' ) as f: SCREAMING_SNAKE_CASE_ : Optional[int] = yaml.safe_load(f.read() ) # Get to the API doc SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 while content[api_idx]["title"] != "API": api_idx += 1 SCREAMING_SNAKE_CASE_ : List[str] = content[api_idx]['sections'] # Then to the model doc SCREAMING_SNAKE_CASE_ : Tuple = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 SCREAMING_SNAKE_CASE_ : Optional[Any] = api_doc[scheduler_idx]['sections'] SCREAMING_SNAKE_CASE_ : int = clean_doc_toc(a ) SCREAMING_SNAKE_CASE_ : List[str] = False if new_scheduler_doc != scheduler_doc: SCREAMING_SNAKE_CASE_ : Tuple = True if overwrite: SCREAMING_SNAKE_CASE_ : Dict = new_scheduler_doc if diff: if overwrite: SCREAMING_SNAKE_CASE_ : Any = api_doc with open(a , 'w' , encoding='utf-8' ) as f: f.write(yaml.dump(a , allow_unicode=a ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) def A_ ( a=False ): """simple docstring""" with open(a , encoding='utf-8' ) as f: SCREAMING_SNAKE_CASE_ : str = yaml.safe_load(f.read() ) # Get to the API doc SCREAMING_SNAKE_CASE_ : int = 0 while content[api_idx]["title"] != "API": api_idx += 1 SCREAMING_SNAKE_CASE_ : Optional[Any] = content[api_idx]['sections'] # Then to the model doc SCREAMING_SNAKE_CASE_ : int = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 SCREAMING_SNAKE_CASE_ : Any = False SCREAMING_SNAKE_CASE_ : int = api_doc[pipeline_idx]['sections'] SCREAMING_SNAKE_CASE_ : Tuple = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: SCREAMING_SNAKE_CASE_ : Union[str, Any] = pipeline_doc['section'] SCREAMING_SNAKE_CASE_ : str = clean_doc_toc(a ) if overwrite: SCREAMING_SNAKE_CASE_ : Optional[int] = new_sub_pipeline_doc new_pipeline_docs.append(a ) # sort overall pipeline doc SCREAMING_SNAKE_CASE_ : Dict = clean_doc_toc(a ) if new_pipeline_docs != pipeline_docs: SCREAMING_SNAKE_CASE_ : int = True if overwrite: SCREAMING_SNAKE_CASE_ : Optional[Any] = new_pipeline_docs if diff: if overwrite: SCREAMING_SNAKE_CASE_ : Optional[Any] = api_doc with open(a , 'w' , encoding='utf-8' ) as f: f.write(yaml.dump(a , allow_unicode=a ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) if __name__ == "__main__": lowerCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') lowerCAmelCase : List[str] = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
353
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _A ( __magic_name__ , unittest.TestCase): SCREAMING_SNAKE_CASE : Dict = GPTSanJapaneseTokenizer SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : Union[str, Any] = {'''do_clean_text''': False, '''add_prefix_space''': False} def UpperCAmelCase ( self ): """simple docstring""" super().setUp() # fmt: off SCREAMING_SNAKE_CASE_ : Optional[Any] = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on SCREAMING_SNAKE_CASE_ : Optional[Any] = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 SCREAMING_SNAKE_CASE_ : int = {'unk_token': '<unk>'} SCREAMING_SNAKE_CASE_ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) SCREAMING_SNAKE_CASE_ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(_SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase ( self , **_SCREAMING_SNAKE_CASE ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = 'こんにちは、世界。 \nこんばんは、㔺界。😀' SCREAMING_SNAKE_CASE_ : Optional[Any] = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE ): """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_input_output_texts(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Any = tokenizer.decode(_SCREAMING_SNAKE_CASE , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE ) return text, ids def UpperCAmelCase ( self ): """simple docstring""" pass # TODO add if relevant def UpperCAmelCase ( self ): """simple docstring""" pass # TODO add if relevant def UpperCAmelCase ( self ): """simple docstring""" pass # TODO add if relevant def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.get_tokenizer() # Testing tokenization SCREAMING_SNAKE_CASE_ : List[Any] = 'こんにちは、世界。 こんばんは、㔺界。' SCREAMING_SNAKE_CASE_ : List[str] = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] SCREAMING_SNAKE_CASE_ : str = tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Testing conversion to ids without special tokens SCREAMING_SNAKE_CASE_ : Optional[Any] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] SCREAMING_SNAKE_CASE_ : List[str] = tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Testing conversion to ids with special tokens SCREAMING_SNAKE_CASE_ : Optional[Any] = tokens + [tokenizer.unk_token] SCREAMING_SNAKE_CASE_ : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] SCREAMING_SNAKE_CASE_ : str = tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizer() # Testing tokenization SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' SCREAMING_SNAKE_CASE_ : str = 'こんにちは、、、、世界。こんばんは、、、、世界。' SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer.encode(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer.decode(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization SCREAMING_SNAKE_CASE_ : int = 'こんにちは、世界。' SCREAMING_SNAKE_CASE_ : List[str] = 'こんばんは、㔺界。😀' SCREAMING_SNAKE_CASE_ : List[str] = 'こんにちは、世界。こんばんは、世界。😀' SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.encode(prefix_text + input_text ) SCREAMING_SNAKE_CASE_ : int = tokenizer.encode('' , prefix_text=prefix_text + input_text ) SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE , prefix_text=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[str] = tokenizer.decode(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : str = tokenizer.decode(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.decode(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization SCREAMING_SNAKE_CASE_ : Any = 'こんにちは、世界。' SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'こんばんは、㔺界。😀' SCREAMING_SNAKE_CASE_ : int = len(tokenizer.encode(_SCREAMING_SNAKE_CASE ) ) - 2 SCREAMING_SNAKE_CASE_ : Union[str, Any] = len(tokenizer.encode(_SCREAMING_SNAKE_CASE ) ) - 2 SCREAMING_SNAKE_CASE_ : str = [1] + [0] * (len_prefix + len_text + 1) SCREAMING_SNAKE_CASE_ : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] SCREAMING_SNAKE_CASE_ : Dict = [1] + [1] * (len_prefix) + [0] * (len_text + 1) SCREAMING_SNAKE_CASE_ : Any = tokenizer(prefix_text + input_text ).token_type_ids SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids SCREAMING_SNAKE_CASE_ : Tuple = tokenizer(_SCREAMING_SNAKE_CASE , prefix_text=_SCREAMING_SNAKE_CASE ).token_type_ids self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer.encode('あンいワ' ) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.encode('' , prefix_text='あンいワ' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(_SCREAMING_SNAKE_CASE ) , tokenizer.decode(_SCREAMING_SNAKE_CASE ) ) self.assertEqual(tokenizer.decode(_SCREAMING_SNAKE_CASE ) , tokenizer.decode(_SCREAMING_SNAKE_CASE ) ) self.assertNotEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertNotEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) SCREAMING_SNAKE_CASE_ : Tuple = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] SCREAMING_SNAKE_CASE_ : Tuple = tokenizer(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer.batch_encode_plus(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE ) # fmt: off SCREAMING_SNAKE_CASE_ : List[str] = [[3_5993, 8640, 2_5948, 3_5998, 3_0647, 3_5675, 3_5999, 3_5999], [3_5993, 1_0382, 9868, 3_5998, 3_0646, 9459, 3_0646, 3_5675]] SCREAMING_SNAKE_CASE_ : int = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] SCREAMING_SNAKE_CASE_ : Any = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , _SCREAMING_SNAKE_CASE ) self.assertListEqual(x_token.token_type_ids , _SCREAMING_SNAKE_CASE ) self.assertListEqual(x_token.attention_mask , _SCREAMING_SNAKE_CASE ) self.assertListEqual(x_token_a.input_ids , _SCREAMING_SNAKE_CASE ) self.assertListEqual(x_token_a.token_type_ids , _SCREAMING_SNAKE_CASE ) self.assertListEqual(x_token_a.attention_mask , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self ): """simple docstring""" pass def UpperCAmelCase ( self ): """simple docstring""" pass
353
1
import argparse from collections import defaultdict import yaml a : Any = "docs/source/en/_toctree.yml" def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] ): __UpperCAmelCase : List[str] = defaultdict(__lowerCamelCase ) __UpperCAmelCase : List[str] = [] __UpperCAmelCase : Optional[int] = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({"""local""": doc["""local"""], """title""": doc["""title"""]} ) else: new_doc_list.append(__lowerCamelCase ) __UpperCAmelCase : Dict = new_doc_list __UpperCAmelCase : Optional[int] = [key for key, value in counts.items() if value > 1] __UpperCAmelCase : Dict = [] for duplicate_key in duplicates: __UpperCAmelCase : Optional[Any] = list({doc["""title"""] for doc in doc_list if doc["""local"""] == duplicate_key} ) if len(__lowerCamelCase ) > 1: raise ValueError( f"""{duplicate_key} is present several times in the documentation table of content at """ """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if """local""" not in counts or counts[doc["""local"""]] == 1] ) __UpperCAmelCase : int = sorted(__lowerCamelCase , key=lambda __lowerCamelCase : s["title"].lower() ) # "overview" gets special treatment and is always first if len(__lowerCamelCase ) > 1: raise ValueError("""{doc_list} has two 'overview' docs which is not allowed.""" ) overview_doc.extend(__lowerCamelCase ) # Sort return overview_doc def lowerCamelCase__ ( __lowerCamelCase : Optional[Any]=False ): with open(__lowerCamelCase , encoding="""utf-8""" ) as f: __UpperCAmelCase : Tuple = yaml.safe_load(f.read() ) # Get to the API doc __UpperCAmelCase : List[str] = 0 while content[api_idx]["title"] != "API": api_idx += 1 __UpperCAmelCase : Any = content[api_idx]["""sections"""] # Then to the model doc __UpperCAmelCase : Optional[int] = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 __UpperCAmelCase : str = api_doc[scheduler_idx]["""sections"""] __UpperCAmelCase : Dict = clean_doc_toc(__lowerCamelCase ) __UpperCAmelCase : List[str] = False if new_scheduler_doc != scheduler_doc: __UpperCAmelCase : Optional[int] = True if overwrite: __UpperCAmelCase : int = new_scheduler_doc if diff: if overwrite: __UpperCAmelCase : str = api_doc with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__lowerCamelCase , allow_unicode=__lowerCamelCase ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) def lowerCamelCase__ ( __lowerCamelCase : List[Any]=False ): with open(__lowerCamelCase , encoding="""utf-8""" ) as f: __UpperCAmelCase : Union[str, Any] = yaml.safe_load(f.read() ) # Get to the API doc __UpperCAmelCase : Optional[int] = 0 while content[api_idx]["title"] != "API": api_idx += 1 __UpperCAmelCase : Tuple = content[api_idx]["""sections"""] # Then to the model doc __UpperCAmelCase : Optional[Any] = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 __UpperCAmelCase : Tuple = False __UpperCAmelCase : int = api_doc[pipeline_idx]["""sections"""] __UpperCAmelCase : Optional[Any] = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: __UpperCAmelCase : Optional[Any] = pipeline_doc["""section"""] __UpperCAmelCase : List[Any] = clean_doc_toc(__lowerCamelCase ) if overwrite: __UpperCAmelCase : List[str] = new_sub_pipeline_doc new_pipeline_docs.append(__lowerCamelCase ) # sort overall pipeline doc __UpperCAmelCase : Any = clean_doc_toc(__lowerCamelCase ) if new_pipeline_docs != pipeline_docs: __UpperCAmelCase : Tuple = True if overwrite: __UpperCAmelCase : Any = new_pipeline_docs if diff: if overwrite: __UpperCAmelCase : List[Any] = api_doc with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__lowerCamelCase , allow_unicode=__lowerCamelCase ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") a : Tuple = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
63
'''simple docstring''' import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def _lowercase ( lowerCamelCase__ : Optional[int], lowerCamelCase__ : Any, lowerCamelCase__ : Tuple=[] ): _a = size[0] - overlap_pixels * 2 _a = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels _a = np.ones((size_y, size_x), dtype=np.uinta ) * 255 _a = np.pad(lowerCamelCase__, mode="linear_ramp", pad_width=lowerCamelCase__, end_values=0 ) if "l" in remove_borders: _a = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: _a = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: _a = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: _a = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def _lowercase ( lowerCamelCase__ : Optional[int], lowerCamelCase__ : Dict, lowerCamelCase__ : Dict ): return max(lowerCamelCase__, min(lowerCamelCase__, lowerCamelCase__ ) ) def _lowercase ( lowerCamelCase__ : [int], lowerCamelCase__ : [int], lowerCamelCase__ : [int] ): return ( clamp(rect[0], min[0], max[0] ), clamp(rect[1], min[1], max[1] ), clamp(rect[2], min[0], max[0] ), clamp(rect[3], min[1], max[1] ), ) def _lowercase ( lowerCamelCase__ : [int], lowerCamelCase__ : int, lowerCamelCase__ : [int] ): _a = list(lowerCamelCase__ ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap _a = clamp_rect(lowerCamelCase__, [0, 0], [image_size[0], image_size[1]] ) return rect def _lowercase ( lowerCamelCase__ : List[str], lowerCamelCase__ : List[str], lowerCamelCase__ : List[str], lowerCamelCase__ : List[str] ): _a = Image.new("RGB", (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]), Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ), (0, 0), ) result.paste(lowerCamelCase__, (original_slice, 0) ) return result def _lowercase ( lowerCamelCase__ : Optional[int], lowerCamelCase__ : Optional[int] ): _a = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) _a = tile.crop(lowerCamelCase__ ) return tile def _lowercase ( lowerCamelCase__ : Any, lowerCamelCase__ : Tuple ): _a = n % d return n - divisor class A ( a ): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = 3_5_0 , ) -> int: super().__init__( vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , unet=snake_case_ , low_res_scheduler=snake_case_ , scheduler=snake_case_ , max_noise_level=snake_case_ , ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> List[Any]: torch.manual_seed(0 ) _a = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) _a = add_overlap_rect(snake_case_ , snake_case_ , image.size ) _a = image.crop(snake_case_ ) _a = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] _a = translated_slice_x - (original_image_slice / 2) _a = max(0 , snake_case_ ) _a = squeeze_tile(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) _a = to_input.size _a = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) _a = super(snake_case_ , self ).__call__(image=snake_case_ , **snake_case_ ).images[0] _a = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) _a = unsqueeze_tile(snake_case_ , snake_case_ ) _a = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) _a = [] if x == 0: remove_borders.append("l" ) elif crop_rect[2] == image.size[0]: remove_borders.append("r" ) if y == 0: remove_borders.append("t" ) elif crop_rect[3] == image.size[1]: remove_borders.append("b" ) _a = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=snake_case_ ) , mode="L" , ) final_image.paste( snake_case_ , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , snake_case_ ) @torch.no_grad() def __call__( self , snake_case_ , snake_case_ , snake_case_ = 7_5 , snake_case_ = 9.0 , snake_case_ = 5_0 , snake_case_ = None , snake_case_ = 1 , snake_case_ = 0.0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = 1 , snake_case_ = 1_2_8 , snake_case_ = 3_2 , snake_case_ = 3_2 , ) -> List[str]: _a = Image.new("RGB" , (image.size[0] * 4, image.size[1] * 4) ) _a = math.ceil(image.size[0] / tile_size ) _a = math.ceil(image.size[1] / tile_size ) _a = tcx * tcy _a = 0 for y in range(snake_case_ ): for x in range(snake_case_ ): self._process_tile( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , prompt=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , noise_level=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , ) current_count += 1 if callback is not None: callback({"progress": current_count / total_tile_count, "image": final_image} ) return final_image def _lowercase ( ): # Run a demo _a = "stabilityai/stable-diffusion-x4-upscaler" _a = StableDiffusionTiledUpscalePipeline.from_pretrained(lowerCamelCase__, revision="fp16", torch_dtype=torch.floataa ) _a = pipe.to("cuda" ) _a = Image.open("../../docs/source/imgs/diffusers_library.jpg" ) def callback(lowerCamelCase__ : Dict ): print(F'''progress: {obj['progress']:.4f}''' ) obj["image"].save("diffusers_library_progress.jpg" ) _a = pipe(image=lowerCamelCase__, prompt="Black font, white background, vector", noise_level=40, callback=lowerCamelCase__ ) final_image.save("diffusers_library.jpg" ) if __name__ == "__main__": main()
131
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __magic_name__ , unittest.TestCase ): _a = KandinskyVaaInpaintPipeline _a = ["""image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] _a = [ """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] _a = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] _a = False @property def UpperCamelCase__ ( self ) -> int: return 32 @property def UpperCamelCase__ ( self ) -> Any: return 32 @property def UpperCamelCase__ ( self ) -> int: return self.time_input_dim @property def UpperCamelCase__ ( self ) -> int: return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ) -> List[str]: return 100 @property def UpperCamelCase__ ( self ) -> int: torch.manual_seed(0 ) __a = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } __a = UNetaDConditionModel(**UpperCamelCase ) return model @property def UpperCamelCase__ ( self ) -> Optional[int]: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCamelCase__ ( self ) -> str: torch.manual_seed(0 ) __a = VQModel(**self.dummy_movq_kwargs ) return model def UpperCamelCase__ ( self ) -> List[str]: __a = self.dummy_unet __a = self.dummy_movq __a = DDIMScheduler( num_train_timesteps=1000 , beta_schedule='linear' , beta_start=0.00_085 , beta_end=0.012 , clip_sample=UpperCamelCase , set_alpha_to_one=UpperCamelCase , steps_offset=1 , prediction_type='epsilon' , thresholding=UpperCamelCase , ) __a = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def UpperCamelCase__ ( self , UpperCamelCase , UpperCamelCase=0 ) -> str: __a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(UpperCamelCase ) ).to(UpperCamelCase ) __a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( UpperCamelCase ) # create init_image __a = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCamelCase ) ).to(UpperCamelCase ) __a = image.cpu().permute(0 , 2 , 3 , 1 )[0] __a = Image.fromarray(np.uinta(UpperCamelCase ) ).convert('RGB' ).resize((256, 256) ) # create mask __a = np.ones((64, 64) , dtype=np.floataa ) __a = 0 if str(UpperCamelCase ).startswith('mps' ): __a = torch.manual_seed(UpperCamelCase ) else: __a = torch.Generator(device=UpperCamelCase ).manual_seed(UpperCamelCase ) __a = { 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def UpperCamelCase__ ( self ) -> str: __a = 'cpu' __a = self.get_dummy_components() __a = self.pipeline_class(**UpperCamelCase ) __a = pipe.to(UpperCamelCase ) pipe.set_progress_bar_config(disable=UpperCamelCase ) __a = pipe(**self.get_dummy_inputs(UpperCamelCase ) ) __a = output.images __a = pipe( **self.get_dummy_inputs(UpperCamelCase ) , return_dict=UpperCamelCase , )[0] __a = image[0, -3:, -3:, -1] __a = image_from_tuple[0, -3:, -3:, -1] print(f"image.shape {image.shape}" ) assert image.shape == (1, 64, 64, 3) __a = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" def UpperCamelCase__ ( self ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def UpperCamelCase__ ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ) -> Tuple: __a = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy' ) __a = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) __a = np.ones((768, 768) , dtype=np.floataa ) __a = 0 __a = 'a hat' __a = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(UpperCamelCase ) __a = KandinskyVaaInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder-inpaint' , torch_dtype=torch.floataa ) __a = pipeline.to(UpperCamelCase ) pipeline.set_progress_bar_config(disable=UpperCamelCase ) __a = torch.Generator(device='cpu' ).manual_seed(0 ) __a , __a = pipe_prior( UpperCamelCase , generator=UpperCamelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() __a = pipeline( image=UpperCamelCase , mask_image=UpperCamelCase , image_embeds=UpperCamelCase , negative_image_embeds=UpperCamelCase , generator=UpperCamelCase , num_inference_steps=100 , height=768 , width=768 , output_type='np' , ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(UpperCamelCase , UpperCamelCase )
720
'''simple docstring''' import os def SCREAMING_SNAKE_CASE ( ): with open(os.path.dirname(a_ ) + '/grid.txt' ) as f: __a = [] # noqa: E741 for _ in range(20 ): l.append([int(a_ ) for x in f.readline().split()] ) __a = 0 # right for i in range(20 ): for j in range(17 ): __a = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: __a = temp # down for i in range(17 ): for j in range(20 ): __a = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: __a = temp # diagonal 1 for i in range(17 ): for j in range(17 ): __a = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: __a = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): __a = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: __a = temp return maximum if __name__ == "__main__": print(solution())
490
0
from __future__ import annotations import math def UpperCamelCase_ ( __a , __a , __a , __a , __a ) -> int: if depth < 0: raise ValueError("Depth cannot be less than 0" ) if len(__a ) == 0: raise ValueError("Scores cannot be empty" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __a , __a , __a ) , minimax(depth + 1 , node_index * 2 + 1 , __a , __a , __a ) , ) return min( minimax(depth + 1 , node_index * 2 , __a , __a , __a ) , minimax(depth + 1 , node_index * 2 + 1 , __a , __a , __a ) , ) def UpperCamelCase_ ( ) -> None: a__ : Optional[Any] = [90, 23, 6, 33, 21, 65, 123, 34_423] a__ : Optional[int] = math.log(len(__a ) , 2 ) print("Optimal value : " , end="" ) print(minimax(0 , 0 , __a , __a , __a ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
37
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch _lowercase: str = '''sshleifer/bart-tiny-random''' _lowercase: Union[str, Any] = '''patrickvonplaten/t5-tiny-random''' @require_torch class lowerCamelCase__ ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): return AutoConfig.from_pretrained(lowercase__ ) def SCREAMING_SNAKE_CASE__ ( self : int ): _lowerCAmelCase , *_lowerCAmelCase = create_student_by_copying_alternating_layers(lowercase__ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): _lowerCAmelCase , *_lowerCAmelCase = create_student_by_copying_alternating_layers(lowercase__ , tempfile.mkdtemp() , e=1 , d=lowercase__ ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): _lowerCAmelCase , *_lowerCAmelCase = create_student_by_copying_alternating_layers(lowercase__ , tempfile.mkdtemp() , e=1 , d=lowercase__ ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def SCREAMING_SNAKE_CASE__ ( self : str ): _lowerCAmelCase , *_lowerCAmelCase = create_student_by_copying_alternating_layers(lowercase__ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): with self.assertRaises(lowercase__ ): create_student_by_copying_alternating_layers(lowercase__ , tempfile.mkdtemp() , e=lowercase__ , d=lowercase__ )
192
0
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins _lowerCAmelCase = ['''tests.fixtures.files''', '''tests.fixtures.hub''', '''tests.fixtures.fsspec'''] def __lowerCAmelCase ( snake_case__ , snake_case__ ) -> Optional[int]: # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ["integration", "unit"] ): continue item.add_marker(pytest.mark.unit ) def __lowerCAmelCase ( snake_case__ ) -> str: config.addinivalue_line("markers" , "torchaudio_latest: mark test to run with torchaudio>=0.12" ) @pytest.fixture(autouse=__UpperCamelCase ) def __lowerCAmelCase ( snake_case__ , snake_case__ ) -> Union[str, Any]: # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? __UpperCamelCase : Optional[int] = tmp_path_factory.getbasetemp() / "cache" __UpperCamelCase : Union[str, Any] = test_hf_cache_home / "datasets" __UpperCamelCase : List[Any] = test_hf_cache_home / "metrics" __UpperCamelCase : List[str] = test_hf_cache_home / "modules" monkeypatch.setattr("datasets.config.HF_DATASETS_CACHE" , str(__UpperCamelCase ) ) monkeypatch.setattr("datasets.config.HF_METRICS_CACHE" , str(__UpperCamelCase ) ) monkeypatch.setattr("datasets.config.HF_MODULES_CACHE" , str(__UpperCamelCase ) ) __UpperCamelCase : Any = test_hf_datasets_cache / "downloads" monkeypatch.setattr("datasets.config.DOWNLOADED_DATASETS_PATH" , str(__UpperCamelCase ) ) __UpperCamelCase : Optional[Any] = test_hf_datasets_cache / "downloads" / "extracted" monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__UpperCamelCase ) ) @pytest.fixture(autouse=__UpperCamelCase , scope="session" ) def __lowerCAmelCase ( ) -> Union[str, Any]: datasets.disable_progress_bar() @pytest.fixture(autouse=__UpperCamelCase ) def __lowerCAmelCase ( snake_case__ ) -> int: # don't take tests into account when counting downloads monkeypatch.setattr("datasets.config.HF_UPDATE_DOWNLOAD_COUNTS" , __UpperCamelCase ) @pytest.fixture def __lowerCAmelCase ( snake_case__ ) -> Any: # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr("sqlalchemy.util.deprecations.SILENCE_UBER_WARNING" , __UpperCamelCase )
708
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' @staticmethod @abstractmethod def a_ (_UpperCAmelCase ) -> List[Any]: raise NotImplementedError() @abstractmethod def a_ (self ) -> Any: raise NotImplementedError()
399
0