code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ : Optional[Any] = { 'configuration_autoformer': [ 'AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'AutoformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = [ 'AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'AutoformerForPrediction', 'AutoformerModel', 'AutoformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys snake_case_ : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1 , UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : int = 1 _UpperCamelCase : Union[str, Any] = 0 for divide_by_number in range(UpperCAmelCase_ , digit + 1 ): _UpperCamelCase : list[int] = [] _UpperCamelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = len(UpperCAmelCase_ ) _UpperCamelCase : List[Any] = divide_by_number else: has_been_divided.append(UpperCAmelCase_ ) _UpperCamelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
83
1
'''simple docstring''' from __future__ import annotations def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Union[str, Any] = 0.00 _UpperCamelCase : Any = 0 for resistor in resistors: if resistor <= 0: _UpperCamelCase : Optional[int] = f'Resistor at index {index} has a negative or zero value!' raise ValueError(UpperCAmelCase_ ) first_sum += 1 / float(UpperCAmelCase_ ) index += 1 return 1 / first_sum def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = 0.00 _UpperCamelCase : Tuple = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _UpperCamelCase : Optional[Any] = f'Resistor at index {index} has a negative value!' raise ValueError(UpperCAmelCase_ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
83
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if num < 0: return False _UpperCamelCase : int = num _UpperCamelCase : int = 0 while num > 0: _UpperCamelCase : str = rev_num * 1_0 + (num % 1_0) num //= 1_0 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
83
1
'''simple docstring''' import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # Initialise PyTorch model _UpperCamelCase : Union[str, Any] = FunnelConfig.from_json_file(UpperCAmelCase_ ) print(f'Building PyTorch model from configuration: {config}' ) _UpperCamelCase : Tuple = FunnelBaseModel(UpperCAmelCase_ ) if base_model else FunnelModel(UpperCAmelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_funnel(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , UpperCAmelCase_ ) if __name__ == "__main__": snake_case_ : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--base_model', action='store_true', help='Whether you want just the base model (no decoder) or not.' ) snake_case_ : Union[str, Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
83
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = abs(UpperCAmelCase_ ) _UpperCamelCase : int = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = abs(UpperCAmelCase_ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def A__ ( UpperCAmelCase_ ): return sum(int(UpperCAmelCase_ ) for c in str(abs(UpperCAmelCase_ ) ) ) def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) -> None: _UpperCamelCase : str = f'{func.__name__}({value})' _UpperCamelCase : Tuple = timeit(f'__main__.{call}' , setup='import __main__' ) print(f'{call:56} = {func(UpperCAmelCase_ )} -- {timing:.4f} seconds' ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
83
1
'''simple docstring''' from __future__ import annotations class lowercase__ : def __init__( self : Optional[int] ,lowerCamelCase__ : Union[str, Any]=None ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = data _UpperCamelCase : Union[str, Any] = None def __repr__( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[Any] = [] _UpperCamelCase : Optional[int] = self while temp: string_rep.append(F'{temp.data}' ) _UpperCamelCase : int = temp.next return "->".join(lowerCamelCase__ ) def A__ ( UpperCAmelCase_ ): if not elements_list: raise Exception('The Elements List is empty' ) _UpperCamelCase : Dict = Node(elements_list[0] ) for i in range(1 , len(UpperCAmelCase_ ) ): _UpperCamelCase : Optional[Any] = Node(elements_list[i] ) _UpperCamelCase : List[Any] = current.next return head def A__ ( UpperCAmelCase_ ): if head_node is not None and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): print_reverse(head_node.next ) print(head_node.data ) def A__ ( ): from doctest import testmod testmod() _UpperCamelCase : int = make_linked_list([1_4, 5_2, 1_4, 1_2, 4_3] ) print('Linked List:' ) print(UpperCAmelCase_ ) print('Elements in Reverse:' ) print_reverse(UpperCAmelCase_ ) if __name__ == "__main__": main()
83
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
1
'''simple docstring''' import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lowercase__ ( unittest.TestCase ): @property def UpperCamelCase_ ( self : int ): '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : Optional[int] = self.dummy_uncond_unet _UpperCamelCase : List[Any] = ScoreSdeVeScheduler() _UpperCamelCase : Optional[int] = ScoreSdeVePipeline(unet=lowerCamelCase__ ,scheduler=lowerCamelCase__ ) sde_ve.to(lowerCamelCase__ ) sde_ve.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = torch.manual_seed(0 ) _UpperCamelCase : str = sde_ve(num_inference_steps=2 ,output_type='numpy' ,generator=lowerCamelCase__ ).images _UpperCamelCase : Dict = torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = sde_ve(num_inference_steps=2 ,output_type='numpy' ,generator=lowerCamelCase__ ,return_dict=lowerCamelCase__ )[ 0 ] _UpperCamelCase : Optional[Any] = image[0, -3:, -3:, -1] _UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase : List[str] = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : List[str] = 'google/ncsnpp-church-256' _UpperCamelCase : Union[str, Any] = UNetaDModel.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : List[str] = ScoreSdeVeScheduler.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : int = ScoreSdeVePipeline(unet=lowerCamelCase__ ,scheduler=lowerCamelCase__ ) sde_ve.to(lowerCamelCase__ ) sde_ve.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = torch.manual_seed(0 ) _UpperCamelCase : int = sde_ve(num_inference_steps=10 ,output_type='numpy' ,generator=lowerCamelCase__ ).images _UpperCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase : int = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
83
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
1
'''simple docstring''' import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class lowercase__ : def __init__( self : Optional[int] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): '''simple docstring''' if dst_width < 0 or dst_height < 0: raise ValueError('Destination width/height should be > 0' ) _UpperCamelCase : Optional[Any] = img _UpperCamelCase : List[str] = img.shape[1] _UpperCamelCase : Optional[Any] = img.shape[0] _UpperCamelCase : str = dst_width _UpperCamelCase : Any = dst_height _UpperCamelCase : Any = self.src_w / self.dst_w _UpperCamelCase : Union[str, Any] = self.src_h / self.dst_h _UpperCamelCase : List[str] = ( np.ones((self.dst_h, self.dst_w, 3) ,np.uinta ) * 255 ) def UpperCamelCase_ ( self : int ): '''simple docstring''' for i in range(self.dst_h ): for j in range(self.dst_w ): _UpperCamelCase : Optional[Any] = self.img[self.get_y(lowerCamelCase__ )][self.get_x(lowerCamelCase__ )] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : int ): '''simple docstring''' return int(self.ratio_x * x ) def UpperCamelCase_ ( self : str ,lowerCamelCase__ : int ): '''simple docstring''' return int(self.ratio_y * y ) if __name__ == "__main__": snake_case_ , snake_case_ : List[str] = 800, 600 snake_case_ : Optional[Any] = imread('image_data/lena.jpg', 1) snake_case_ : Dict = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F"""Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}""", n.output ) waitKey(0) destroyAllWindows()
83
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
1
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ : Optional[Any] = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Optional[int] = [ 'XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaXLForCausalLM', 'XLMRobertaXLForMaskedLM', 'XLMRobertaXLForMultipleChoice', 'XLMRobertaXLForQuestionAnswering', 'XLMRobertaXLForSequenceClassification', 'XLMRobertaXLForTokenClassification', 'XLMRobertaXLModel', 'XLMRobertaXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys snake_case_ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure)
83
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
1
'''simple docstring''' import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase_ ( self : Optional[int] ,lowerCamelCase__ : Optional[int]=0 ): '''simple docstring''' _UpperCamelCase : List[Any] = np.random.RandomState(lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Tuple = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = self.get_dummy_inputs() _UpperCamelCase : Optional[Any] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : str = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Dict = PNDMScheduler.from_config(pipe.scheduler.config ,skip_prk_steps=lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Dict = self.get_dummy_inputs() _UpperCamelCase : str = pipe(**lowerCamelCase__ ).images _UpperCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : Dict = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : List[Any] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = self.get_dummy_inputs() _UpperCamelCase : str = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : List[str] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : str = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.get_dummy_inputs() _UpperCamelCase : Union[str, Any] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : Dict = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Tuple = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Any = self.get_dummy_inputs() _UpperCamelCase : List[Any] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : List[Any] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : List[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : int = self.get_dummy_inputs() _UpperCamelCase : Optional[int] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _UpperCamelCase : Dict = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = self.get_dummy_inputs() _UpperCamelCase : int = 3 * [inputs['prompt']] # forward _UpperCamelCase : Union[str, Any] = pipe(**lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = output.images[0, -3:, -3:, -1] _UpperCamelCase : Dict = self.get_dummy_inputs() _UpperCamelCase : Optional[int] = 3 * [inputs.pop('prompt' )] _UpperCamelCase : Tuple = pipe.tokenizer( lowerCamelCase__ ,padding='max_length' ,max_length=pipe.tokenizer.model_max_length ,truncation=lowerCamelCase__ ,return_tensors='np' ,) _UpperCamelCase : Dict = text_inputs['input_ids'] _UpperCamelCase : Optional[Any] = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] _UpperCamelCase : Tuple = prompt_embeds # forward _UpperCamelCase : Dict = pipe(**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : int = self.get_dummy_inputs() _UpperCamelCase : Dict = 3 * ['this is a negative prompt'] _UpperCamelCase : Optional[int] = negative_prompt _UpperCamelCase : Tuple = 3 * [inputs['prompt']] # forward _UpperCamelCase : Tuple = pipe(**lowerCamelCase__ ) _UpperCamelCase : int = output.images[0, -3:, -3:, -1] _UpperCamelCase : List[Any] = self.get_dummy_inputs() _UpperCamelCase : Optional[int] = 3 * [inputs.pop('prompt' )] _UpperCamelCase : List[str] = [] for p in [prompt, negative_prompt]: _UpperCamelCase : int = pipe.tokenizer( lowerCamelCase__ ,padding='max_length' ,max_length=pipe.tokenizer.model_max_length ,truncation=lowerCamelCase__ ,return_tensors='np' ,) _UpperCamelCase : Optional[Any] = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) _UpperCamelCase , _UpperCamelCase : Optional[Any] = embeds # forward _UpperCamelCase : Dict = pipe(**lowerCamelCase__ ) _UpperCamelCase : int = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class lowercase__ ( unittest.TestCase ): @property def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Dict = ort.SessionOptions() _UpperCamelCase : str = False return options def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' # using the PNDM scheduler by default _UpperCamelCase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' ,revision='onnx' ,safety_checker=lowerCamelCase__ ,feature_extractor=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0 ) _UpperCamelCase : int = sd_pipe([prompt] ,guidance_scale=6.0 ,num_inference_steps=10 ,output_type='np' ) _UpperCamelCase : List[Any] = output.images _UpperCamelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Tuple = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[Any] = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,subfolder='scheduler' ,revision='onnx' ) _UpperCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,revision='onnx' ,scheduler=lowerCamelCase__ ,safety_checker=lowerCamelCase__ ,feature_extractor=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Any = 'open neural network exchange' _UpperCamelCase : List[Any] = np.random.RandomState(0 ) _UpperCamelCase : List[Any] = sd_pipe([prompt] ,guidance_scale=7.5 ,num_inference_steps=10 ,generator=lowerCamelCase__ ,output_type='np' ) _UpperCamelCase : Optional[Any] = output.images _UpperCamelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : Any = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,subfolder='scheduler' ,revision='onnx' ) _UpperCamelCase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,revision='onnx' ,scheduler=lowerCamelCase__ ,safety_checker=lowerCamelCase__ ,feature_extractor=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : List[Any] = 'open neural network exchange' _UpperCamelCase : Dict = np.random.RandomState(0 ) _UpperCamelCase : Union[str, Any] = sd_pipe([prompt] ,guidance_scale=7.5 ,num_inference_steps=10 ,generator=lowerCamelCase__ ,output_type='np' ) _UpperCamelCase : Union[str, Any] = output.images _UpperCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : Dict = 0 def test_callback_fn(lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : np.ndarray ) -> None: _UpperCamelCase : Union[str, Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _UpperCamelCase : Any = latents[0, -3:, -3:, -1] _UpperCamelCase : Dict = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _UpperCamelCase : Union[str, Any] = latents[0, -3:, -3:, -1] _UpperCamelCase : int = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 _UpperCamelCase : Dict = False _UpperCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,revision='onnx' ,safety_checker=lowerCamelCase__ ,feature_extractor=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = 'Andromeda galaxy in a bottle' _UpperCamelCase : Optional[Any] = np.random.RandomState(0 ) pipe( prompt=lowerCamelCase__ ,num_inference_steps=5 ,guidance_scale=7.5 ,generator=lowerCamelCase__ ,callback=lowerCamelCase__ ,callback_steps=1 ,) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' ,revision='onnx' ,safety_checker=lowerCamelCase__ ,feature_extractor=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) assert isinstance(lowerCamelCase__ ,lowerCamelCase__ ) assert pipe.safety_checker is None _UpperCamelCase : List[Any] = pipe('example prompt' ,num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) _UpperCamelCase : Tuple = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None _UpperCamelCase : Union[str, Any] = pipe('example prompt' ,num_inference_steps=2 ).images[0] assert image is not None
83
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase__ ( lowercase ): def __init__( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : str = dataset _UpperCamelCase : Optional[Any] = process _UpperCamelCase : Optional[Any] = params def __len__( self : Tuple ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.dataset[i] _UpperCamelCase : Dict = self.process(lowerCamelCase__ ,**self.params ) return processed class lowercase__ ( lowercase ): def __init__( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int]=None ): '''simple docstring''' _UpperCamelCase : Optional[int] = loader _UpperCamelCase : Tuple = infer _UpperCamelCase : List[str] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _UpperCamelCase : Any = None _UpperCamelCase : Union[str, Any] = loader_batch_size # Internal bookkeeping _UpperCamelCase : Optional[Any] = None _UpperCamelCase : str = None def __len__( self : List[str] ): '''simple docstring''' return len(self.loader ) def __iter__( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = iter(self.loader ) return self def UpperCamelCase_ ( self : Any ): '''simple docstring''' if isinstance(self._loader_batch_data ,torch.Tensor ): # Batch data is simple tensor, just fetch the slice _UpperCamelCase : Union[str, Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _UpperCamelCase : Union[str, Any] = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Convert ModelOutput to tuple first _UpperCamelCase : str = element.to_tuple() if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Dict = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _UpperCamelCase : Optional[int] = None elif isinstance(element[self._loader_batch_index] ,torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] ,np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : Optional[Any] = np.expand_dims(element[self._loader_batch_index] ,0 ) else: # This is typically a list, so no need to `unsqueeze`. _UpperCamelCase : Union[str, Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _UpperCamelCase : Optional[int] = self._loader_batch_data.__class__(lowerCamelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _UpperCamelCase : Tuple = next(self.iterator ) _UpperCamelCase : List[str] = self.infer(lowerCamelCase__ ,**self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : List[Any] = processed else: _UpperCamelCase : List[Any] = list(processed.keys() )[0] _UpperCamelCase : Optional[int] = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : int = len(lowerCamelCase__ ) else: _UpperCamelCase : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : int = observed_batch_size # Setting internal index to unwrap the batch _UpperCamelCase : Dict = processed _UpperCamelCase : str = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase__ ( lowercase ): def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __iter__( self : Dict ): '''simple docstring''' _UpperCamelCase : str = iter(self.loader ) _UpperCamelCase : List[str] = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: _UpperCamelCase : Tuple = self.infer(next(self.iterator ) ,**self.params ) try: # Try to return next item _UpperCamelCase : Optional[Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) _UpperCamelCase : int = next(self.subiterator ) return processed class lowercase__ ( lowercase ): def __iter__( self : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : Dict = self.loader_batch_item() _UpperCamelCase : List[str] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator while not is_last: _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) if self.loader_batch_size is not None: if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : str = processed else: _UpperCamelCase : Any = list(processed.keys() )[0] _UpperCamelCase : Tuple = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Dict = len(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : Any = observed_batch_size _UpperCamelCase : List[Any] = processed _UpperCamelCase : int = 0 while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : List[Any] = self.loader_batch_item() _UpperCamelCase : Optional[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator else: _UpperCamelCase : Any = processed _UpperCamelCase : List[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) return accumulator class lowercase__ ( lowercase ): def __init__( self : Tuple ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : str = key def __len__( self : Dict ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : Optional[Any] = keya _UpperCamelCase : str = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : List[str] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
83
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) snake_case_ : int = { 'configuration_vision_text_dual_encoder': ['VisionTextDualEncoderConfig'], 'processing_vision_text_dual_encoder': ['VisionTextDualEncoderProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : str = ['VisionTextDualEncoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Any = ['FlaxVisionTextDualEncoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Union[str, Any] = ['TFVisionTextDualEncoderModel'] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys snake_case_ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
83
'''simple docstring''' import os from datetime import datetime as dt from github import Github snake_case_ : Any = [ 'good first issue', 'good second issue', 'good difficult issue', 'enhancement', 'new pipeline/model', 'new scheduler', 'wip', ] def A__ ( ): _UpperCamelCase : Tuple = Github(os.environ['GITHUB_TOKEN'] ) _UpperCamelCase : List[Any] = g.get_repo('huggingface/diffusers' ) _UpperCamelCase : List[Any] = repo.get_issues(state='open' ) for issue in open_issues: _UpperCamelCase : Dict = sorted(issue.get_comments() , key=lambda UpperCAmelCase_ : i.created_at , reverse=UpperCAmelCase_ ) _UpperCamelCase : List[str] = comments[0] if len(UpperCAmelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='closed' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='open' ) issue.remove_from_labels('stale' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) issue.add_to_labels('stale' ) if __name__ == "__main__": main()
83
1
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class lowercase__ : def __init__( self : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : str=13 ,lowerCamelCase__ : Any=7 ,lowerCamelCase__ : int=False ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : int=False ,lowerCamelCase__ : str=True ,lowerCamelCase__ : List[Any]=33 ,lowerCamelCase__ : List[Any]=32 ,lowerCamelCase__ : str=5 ,lowerCamelCase__ : str=4 ,lowerCamelCase__ : List[Any]=37 ,lowerCamelCase__ : Union[str, Any]="gelu" ,lowerCamelCase__ : List[str]=0.1 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : str=512 ,lowerCamelCase__ : Optional[int]=16 ,lowerCamelCase__ : Optional[Any]=2 ,lowerCamelCase__ : str=0.0_2 ,lowerCamelCase__ : Any=3 ,lowerCamelCase__ : Union[str, Any]=4 ,lowerCamelCase__ : Tuple=None ,): '''simple docstring''' _UpperCamelCase : str = parent _UpperCamelCase : int = batch_size _UpperCamelCase : Dict = seq_length _UpperCamelCase : Dict = is_training _UpperCamelCase : Optional[int] = use_input_mask _UpperCamelCase : Optional[int] = use_token_type_ids _UpperCamelCase : Tuple = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : Dict = hidden_size _UpperCamelCase : Union[str, Any] = num_hidden_layers _UpperCamelCase : int = num_attention_heads _UpperCamelCase : Dict = intermediate_size _UpperCamelCase : str = hidden_act _UpperCamelCase : int = hidden_dropout_prob _UpperCamelCase : List[Any] = attention_probs_dropout_prob _UpperCamelCase : List[Any] = max_position_embeddings _UpperCamelCase : Tuple = type_vocab_size _UpperCamelCase : List[str] = type_sequence_label_size _UpperCamelCase : str = initializer_range _UpperCamelCase : Union[str, Any] = num_labels _UpperCamelCase : Union[str, Any] = num_choices _UpperCamelCase : Any = scope def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Any = None if self.use_input_mask: _UpperCamelCase : int = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : List[Any] = None _UpperCamelCase : List[str] = None _UpperCamelCase : List[Any] = None if self.use_labels: _UpperCamelCase : Tuple = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) _UpperCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) _UpperCamelCase : Any = ids_tensor([self.batch_size] ,self.num_choices ) _UpperCamelCase : Union[str, Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,pad_token_id=1 ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Dict ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = EsmModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCamelCase : Union[str, Any] = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ ) _UpperCamelCase : Any = model(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : List[str] = EsmForMaskedLM(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCamelCase : Tuple = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self : int ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' _UpperCamelCase : Any = self.num_labels _UpperCamelCase : Optional[int] = EsmForTokenClassification(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCamelCase : Union[str, Any] = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : str = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) : Optional[Any] = config_and_inputs _UpperCamelCase : int = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowercase__ ( lowercase , lowercase , unittest.TestCase ): lowercase__ = False lowercase__ = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) lowercase__ = () lowercase__ = ( { """feature-extraction""": EsmModel, """fill-mask""": EsmForMaskedLM, """text-classification""": EsmForSequenceClassification, """token-classification""": EsmForTokenClassification, """zero-shot""": EsmForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Dict = EsmModelTester(self ) _UpperCamelCase : List[str] = ConfigTester(self ,config_class=lowerCamelCase__ ,hidden_size=37 ) def UpperCamelCase_ ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase : int = type self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase : Optional[int] = EsmModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs()[0] _UpperCamelCase : Optional[int] = EsmEmbeddings(config=lowerCamelCase__ ) _UpperCamelCase : List[str] = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _UpperCamelCase : Optional[Any] = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _UpperCamelCase : int = create_position_ids_from_input_ids(lowerCamelCase__ ,model.padding_idx ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(lowerCamelCase__ ,lowerCamelCase__ ) ) ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs()[0] _UpperCamelCase : List[str] = EsmEmbeddings(config=lowerCamelCase__ ) _UpperCamelCase : List[Any] = torch.empty(2 ,4 ,30 ) _UpperCamelCase : Optional[int] = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _UpperCamelCase : Union[str, Any] = torch.as_tensor([expected_single_positions, expected_single_positions] ) _UpperCamelCase : Dict = embeddings.create_position_ids_from_inputs_embeds(lowerCamelCase__ ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(lowerCamelCase__ ,lowerCamelCase__ ) ) ) @unittest.skip('Esm does not support embedding resizing' ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip('Esm does not support embedding resizing' ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' pass @require_torch class lowercase__ ( lowercase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' with torch.no_grad(): _UpperCamelCase : Any = EsmForMaskedLM.from_pretrained('facebook/esm2_t6_8M_UR50D' ) model.eval() _UpperCamelCase : int = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _UpperCamelCase : List[str] = model(lowerCamelCase__ )[0] _UpperCamelCase : List[Any] = 33 _UpperCamelCase : Dict = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Optional[int] = torch.tensor( [[[8.9_2_1_5, -1_0.5_8_9_8, -6.4_6_7_1], [-6.3_9_6_7, -1_3.9_1_1_4, -1.1_2_1_2], [-7.7_8_1_2, -1_3.9_5_1_6, -3.7_4_0_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,lowerCamelCase__ ,atol=1E-4 ) ) @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' with torch.no_grad(): _UpperCamelCase : Optional[Any] = EsmModel.from_pretrained('facebook/esm2_t6_8M_UR50D' ) model.eval() _UpperCamelCase : str = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _UpperCamelCase : List[Any] = model(lowerCamelCase__ )[0] # compare the actual values for a slice. _UpperCamelCase : Dict = torch.tensor( [[[0.1_4_4_4, 0.5_4_1_3, 0.3_2_4_8], [0.3_0_3_4, 0.0_0_5_3, 0.3_1_0_8], [0.3_2_2_8, -0.2_4_9_9, 0.3_4_1_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,lowerCamelCase__ ,atol=1E-4 ) )
83
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase ) , """Tatoeba directory does not exist.""" ) class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Dict = self.resolver.write_model_card('opus-mt-he-en' ,dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
1
'''simple docstring''' import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase__ ( lowercase ): lowercase__ = (DEISMultistepScheduler,) lowercase__ = (("""num_inference_steps""", 25),) def UpperCamelCase_ ( self : Dict ,**lowerCamelCase__ : Optional[int] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'solver_order': 2, } config.update(**lowerCamelCase__ ) return config def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : Tuple=0 ,**lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = dict(self.forward_default_kwargs ) _UpperCamelCase : int = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) _UpperCamelCase : int = self.dummy_sample _UpperCamelCase : Union[str, Any] = 0.1 * sample _UpperCamelCase : str = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _UpperCamelCase : Optional[Any] = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCamelCase : Optional[int] = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _UpperCamelCase : Optional[Any] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _UpperCamelCase : Tuple = scheduler_class.from_pretrained(lowerCamelCase__ ) new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _UpperCamelCase : str = dummy_past_residuals[: new_scheduler.config.solver_order] _UpperCamelCase , _UpperCamelCase : Any = sample, sample for t in range(lowerCamelCase__ ,time_step + scheduler.config.solver_order + 1 ): _UpperCamelCase : int = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample _UpperCamelCase : Optional[Any] = new_scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' pass def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[str]=0 ,**lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = dict(self.forward_default_kwargs ) _UpperCamelCase : List[Any] = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) _UpperCamelCase : int = self.dummy_sample _UpperCamelCase : str = 0.1 * sample _UpperCamelCase : Tuple = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _UpperCamelCase : Optional[int] = self.get_scheduler_config() _UpperCamelCase : Tuple = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _UpperCamelCase : List[str] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _UpperCamelCase : Optional[int] = scheduler_class.from_pretrained(lowerCamelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residual (must be after setting timesteps) _UpperCamelCase : Optional[Any] = dummy_past_residuals[: new_scheduler.config.solver_order] _UpperCamelCase : Optional[int] = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample _UpperCamelCase : List[Any] = new_scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int=None ,**lowerCamelCase__ : Any ): '''simple docstring''' if scheduler is None: _UpperCamelCase : Any = self.scheduler_classes[0] _UpperCamelCase : Optional[Any] = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCamelCase : str = scheduler_class(**lowerCamelCase__ ) _UpperCamelCase : List[Any] = self.scheduler_classes[0] _UpperCamelCase : str = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCamelCase : int = scheduler_class(**lowerCamelCase__ ) _UpperCamelCase : Dict = 10 _UpperCamelCase : Optional[int] = self.dummy_model() _UpperCamelCase : Tuple = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCamelCase : Dict = model(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : List[str] = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample return sample def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : Any = dict(self.forward_default_kwargs ) _UpperCamelCase : Dict = kwargs.pop('num_inference_steps' ,lowerCamelCase__ ) for scheduler_class in self.scheduler_classes: _UpperCamelCase : Union[str, Any] = self.get_scheduler_config() _UpperCamelCase : Optional[int] = scheduler_class(**lowerCamelCase__ ) _UpperCamelCase : str = self.dummy_sample _UpperCamelCase : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowerCamelCase__ ,'set_timesteps' ): scheduler.set_timesteps(lowerCamelCase__ ) elif num_inference_steps is not None and not hasattr(lowerCamelCase__ ,'set_timesteps' ): _UpperCamelCase : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _UpperCamelCase : Tuple = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] _UpperCamelCase : List[str] = dummy_past_residuals[: scheduler.config.solver_order] _UpperCamelCase : Union[str, Any] = scheduler.timesteps[5] _UpperCamelCase : int = scheduler.timesteps[6] _UpperCamelCase : Tuple = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample _UpperCamelCase : str = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # make sure that iterating over schedulers with same config names gives same results # for defaults _UpperCamelCase : List[str] = DEISMultistepScheduler(**self.get_scheduler_config() ) _UpperCamelCase : Any = self.full_loop(scheduler=lowerCamelCase__ ) _UpperCamelCase : List[Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 _UpperCamelCase : str = DPMSolverSinglestepScheduler.from_config(scheduler.config ) _UpperCamelCase : List[str] = DPMSolverMultistepScheduler.from_config(scheduler.config ) _UpperCamelCase : Union[str, Any] = UniPCMultistepScheduler.from_config(scheduler.config ) _UpperCamelCase : List[str] = DEISMultistepScheduler.from_config(scheduler.config ) _UpperCamelCase : Dict = self.full_loop(scheduler=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.check_over_configs(thresholding=lowerCamelCase__ ) for order in [1, 2, 3]: for solver_type in ["logrho"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCamelCase__ ,prediction_type=lowerCamelCase__ ,sample_max_value=lowerCamelCase__ ,algorithm_type='deis' ,solver_order=lowerCamelCase__ ,solver_type=lowerCamelCase__ ,) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' for algorithm_type in ["deis"]: for solver_type in ["logrho"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCamelCase__ ,solver_type=lowerCamelCase__ ,prediction_type=lowerCamelCase__ ,algorithm_type=lowerCamelCase__ ,) _UpperCamelCase : int = self.full_loop( solver_order=lowerCamelCase__ ,solver_type=lowerCamelCase__ ,prediction_type=lowerCamelCase__ ,algorithm_type=lowerCamelCase__ ,) assert not torch.isnan(lowerCamelCase__ ).any(), "Samples have nan numbers" def UpperCamelCase_ ( self : int ): '''simple docstring''' self.check_over_configs(lower_order_final=lowerCamelCase__ ) self.check_over_configs(lower_order_final=lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCamelCase__ ,time_step=0 ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : str = self.full_loop() _UpperCamelCase : Tuple = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.full_loop(prediction_type='v_prediction' ) _UpperCamelCase : Tuple = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.0_9_1 ) < 1E-3 def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[int] = self.scheduler_classes[0] _UpperCamelCase : List[Any] = self.get_scheduler_config(thresholding=lowerCamelCase__ ,dynamic_thresholding_ratio=0 ) _UpperCamelCase : Any = scheduler_class(**lowerCamelCase__ ) _UpperCamelCase : Dict = 10 _UpperCamelCase : List[Any] = self.dummy_model() _UpperCamelCase : Any = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCamelCase : Optional[Any] = model(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : List[Any] = scheduler.step(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ).prev_sample assert sample.dtype == torch.floataa
83
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
1
'''simple docstring''' import re def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = re.compile(R'^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$' ) if match := re.search(UpperCAmelCase_ , UpperCAmelCase_ ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
83
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : Dict = 3 _UpperCamelCase : Any = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 1_5 == 0: result -= a a += 1 return result if __name__ == "__main__": print(F"""{solution() = }""")
83
1
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake snake_case_ : List[Any] = numpy.array([0, 0]) snake_case_ : Optional[Any] = numpy.array([0.5, 0.8_66_02_54]) snake_case_ : int = numpy.array([1, 0]) snake_case_ : Optional[int] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = initial_vectors for _ in range(UpperCAmelCase_ ): _UpperCamelCase : Any = iteration_step(UpperCAmelCase_ ) return vectors def A__ ( UpperCAmelCase_ ): _UpperCamelCase : int = [] for i, start_vector in enumerate(vectors[:-1] ): _UpperCamelCase : Union[str, Any] = vectors[i + 1] new_vectors.append(UpperCAmelCase_ ) _UpperCamelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 6_0 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Dict = numpy.radians(UpperCAmelCase_ ) _UpperCamelCase , _UpperCamelCase : Any = numpy.cos(UpperCAmelCase_ ), numpy.sin(UpperCAmelCase_ ) _UpperCamelCase : int = numpy.array(((c, -s), (s, c)) ) return numpy.dot(UpperCAmelCase_ , UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Any = plt.gca() axes.set_aspect('equal' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _UpperCamelCase , _UpperCamelCase : Optional[Any] = zip(*UpperCAmelCase_ ) plt.plot(UpperCAmelCase_ , UpperCAmelCase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() snake_case_ : Any = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
83
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
1
'''simple docstring''' import unittest import torch from torch import nn from diffusers.models.activations import get_activation class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = get_activation('swish' ) self.assertIsInstance(lowerCamelCase__ ,nn.SiLU ) self.assertEqual(act(torch.tensor(-100 ,dtype=torch.floataa ) ).item() ,0 ) self.assertNotEqual(act(torch.tensor(-1 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(0 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(20 ,dtype=torch.floataa ) ).item() ,20 ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = get_activation('silu' ) self.assertIsInstance(lowerCamelCase__ ,nn.SiLU ) self.assertEqual(act(torch.tensor(-100 ,dtype=torch.floataa ) ).item() ,0 ) self.assertNotEqual(act(torch.tensor(-1 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(0 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(20 ,dtype=torch.floataa ) ).item() ,20 ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Dict = get_activation('mish' ) self.assertIsInstance(lowerCamelCase__ ,nn.Mish ) self.assertEqual(act(torch.tensor(-200 ,dtype=torch.floataa ) ).item() ,0 ) self.assertNotEqual(act(torch.tensor(-1 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(0 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(20 ,dtype=torch.floataa ) ).item() ,20 ) def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Dict = get_activation('gelu' ) self.assertIsInstance(lowerCamelCase__ ,nn.GELU ) self.assertEqual(act(torch.tensor(-100 ,dtype=torch.floataa ) ).item() ,0 ) self.assertNotEqual(act(torch.tensor(-1 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(0 ,dtype=torch.floataa ) ).item() ,0 ) self.assertEqual(act(torch.tensor(20 ,dtype=torch.floataa ) ).item() ,20 )
83
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') snake_case_ : Any = logging.getLogger(__name__) @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) lowercase__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class lowercase__ : lowercase__ = field(default=lowercase , metadata={"""help""": """The input training data file (a text file)."""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def UpperCamelCase_ ( self : str ): '''simple docstring''' if self.train_file is not None: _UpperCamelCase : List[Any] = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _UpperCamelCase : Union[str, Any] = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowercase__ : lowercase__ = 42 lowercase__ = True lowercase__ = None lowercase__ = None def __call__( self : Optional[Any] ,lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = 'label' if 'label' in features[0].keys() else 'labels' _UpperCamelCase : List[Any] = [feature.pop(lowerCamelCase__ ) for feature in features] _UpperCamelCase : Dict = len(lowerCamelCase__ ) _UpperCamelCase : List[str] = len(features[0]['input_ids'] ) _UpperCamelCase : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(lowerCamelCase__ )] for feature in features ] _UpperCamelCase : str = list(chain(*lowerCamelCase__ ) ) _UpperCamelCase : Tuple = self.tokenizer.pad( lowerCamelCase__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='pt' ,) # Un-flatten _UpperCamelCase : str = {k: v.view(lowerCamelCase__ ,lowerCamelCase__ ,-1 ) for k, v in batch.items()} # Add back labels _UpperCamelCase : Optional[int] = torch.tensor(lowerCamelCase__ ,dtype=torch.intaa ) return batch def A__ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCamelCase : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Optional[int] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , UpperCAmelCase_ , UpperCAmelCase_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(UpperCAmelCase_ ) datasets.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCamelCase : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCamelCase : List[str] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _UpperCamelCase : Optional[int] = {} if data_args.train_file is not None: _UpperCamelCase : Tuple = data_args.train_file if data_args.validation_file is not None: _UpperCamelCase : Tuple = data_args.validation_file _UpperCamelCase : Any = data_args.train_file.split('.' )[-1] _UpperCamelCase : Union[str, Any] = load_dataset( UpperCAmelCase_ , data_files=UpperCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _UpperCamelCase : List[str] = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=UpperCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _UpperCamelCase : Any = [f'ending{i}' for i in range(4 )] _UpperCamelCase : int = 'sent1' _UpperCamelCase : List[str] = 'sent2' if data_args.max_seq_length is None: _UpperCamelCase : int = tokenizer.model_max_length if max_seq_length > 1_0_2_4: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) _UpperCamelCase : int = 1_0_2_4 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) _UpperCamelCase : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(UpperCAmelCase_ ): _UpperCamelCase : str = [[context] * 4 for context in examples[context_name]] _UpperCamelCase : Optional[Any] = examples[question_header_name] _UpperCamelCase : Tuple = [ [f'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(UpperCAmelCase_ ) ] # Flatten out _UpperCamelCase : Optional[int] = list(chain(*UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = list(chain(*UpperCAmelCase_ ) ) # Tokenize _UpperCamelCase : Tuple = tokenizer( UpperCAmelCase_ , UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(UpperCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _UpperCamelCase : Optional[Any] = raw_datasets['train'] if data_args.max_train_samples is not None: _UpperCamelCase : Tuple = min(len(UpperCAmelCase_ ) , data_args.max_train_samples ) _UpperCamelCase : Tuple = train_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): _UpperCamelCase : Union[str, Any] = train_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _UpperCamelCase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _UpperCamelCase : Union[str, Any] = min(len(UpperCAmelCase_ ) , data_args.max_eval_samples ) _UpperCamelCase : str = eval_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): _UpperCamelCase : Dict = eval_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _UpperCamelCase : List[Any] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=UpperCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Union[str, Any] = eval_predictions _UpperCamelCase : List[str] = np.argmax(UpperCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer _UpperCamelCase : Optional[int] = Trainer( model=UpperCAmelCase_ , args=UpperCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=UpperCAmelCase_ , data_collator=UpperCAmelCase_ , compute_metrics=UpperCAmelCase_ , ) # Training if training_args.do_train: _UpperCamelCase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _UpperCamelCase : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCamelCase : int = last_checkpoint _UpperCamelCase : List[str] = trainer.train(resume_from_checkpoint=UpperCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload _UpperCamelCase : Union[str, Any] = train_result.metrics _UpperCamelCase : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('train' , UpperCAmelCase_ ) trainer.save_metrics('train' , UpperCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _UpperCamelCase : List[Any] = trainer.evaluate() _UpperCamelCase : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(UpperCAmelCase_ ) _UpperCamelCase : int = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('eval' , UpperCAmelCase_ ) trainer.save_metrics('eval' , UpperCAmelCase_ ) _UpperCamelCase : Optional[int] = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**UpperCAmelCase_ ) else: trainer.create_model_card(**UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
83
1
'''simple docstring''' import sys snake_case_ : Dict = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def A__ ( UpperCAmelCase_ = N ): _UpperCamelCase : Optional[Any] = -sys.maxsize - 1 for i in range(len(UpperCAmelCase_ ) - 1_2 ): _UpperCamelCase : List[Any] = 1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: _UpperCamelCase : Union[str, Any] = product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
83
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) lowercase__ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) lowercase__ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def A__ ( ): _UpperCamelCase : Optional[Any] = HfArgumentParser((ModelArguments,) ) ((_UpperCamelCase) , ) : Optional[int] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _UpperCamelCase : Any = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _UpperCamelCase : List[Any] = True _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : str = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=UpperCAmelCase_ , decoder_config=UpperCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _UpperCamelCase : str = decoder_config.decoder_start_token_id _UpperCamelCase : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _UpperCamelCase : int = decoder_config.bos_token_id if pad_token_id is None: _UpperCamelCase : Dict = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _UpperCamelCase : List[Any] = decoder_config.eos_token_id _UpperCamelCase : Dict = decoder_start_token_id _UpperCamelCase : int = pad_token_id _UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
83
1
'''simple docstring''' def A__ ( UpperCAmelCase_ = 3 , UpperCAmelCase_ = 7 , UpperCAmelCase_ = 1_0_0_0_0_0_0 ): _UpperCamelCase : Optional[Any] = 0 _UpperCamelCase : Tuple = 1 for current_denominator in range(1 , limit + 1 ): _UpperCamelCase : int = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: _UpperCamelCase : List[Any] = current_numerator _UpperCamelCase : List[Any] = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1000000))
83
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy snake_case_ : Dict = logging.get_logger(__name__) class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : float ,**lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : List[Any] = feature_size _UpperCamelCase : Any = sampling_rate _UpperCamelCase : Optional[Any] = padding_value _UpperCamelCase : Union[str, Any] = kwargs.pop('padding_side' ,'right' ) _UpperCamelCase : Dict = kwargs.pop('return_attention_mask' ,lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = True ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,): '''simple docstring''' # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowerCamelCase__ ,(list, tuple) ) and isinstance(processed_features[0] ,(dict, BatchFeature) ): _UpperCamelCase : int = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' F' to this method that includes {self.model_input_names[0]}, but you provided' F' {list(processed_features.keys() )}' ) _UpperCamelCase : List[Any] = processed_features[self.model_input_names[0]] _UpperCamelCase : Dict = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowerCamelCase__ ) == 0: if return_attention_mask: _UpperCamelCase : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch _UpperCamelCase : List[str] = required_input[0] if isinstance(lowerCamelCase__ ,(list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. _UpperCamelCase : List[str] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowerCamelCase__ ): _UpperCamelCase : Dict = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowerCamelCase__ ): _UpperCamelCase : Any = 'tf' elif is_torch_tensor(lowerCamelCase__ ): _UpperCamelCase : Optional[int] = 'pt' elif isinstance(lowerCamelCase__ ,(int, float, list, tuple, np.ndarray) ): _UpperCamelCase : int = 'np' else: raise ValueError( F'type of {first_element} unknown: {type(lowerCamelCase__ )}. ' 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] ,(int, float) ): _UpperCamelCase : Any = to_numpy(lowerCamelCase__ ) else: _UpperCamelCase : Any = [to_numpy(lowerCamelCase__ ) for v in value] # Convert padding_strategy in PaddingStrategy _UpperCamelCase : Optional[int] = self._get_padding_strategies(padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ) _UpperCamelCase : str = processed_features[self.model_input_names[0]] _UpperCamelCase : List[str] = len(lowerCamelCase__ ) if not all(len(lowerCamelCase__ ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) _UpperCamelCase : List[str] = [] for i in range(lowerCamelCase__ ): _UpperCamelCase : List[str] = {k: v[i] for k, v in processed_features.items()} # truncation _UpperCamelCase : List[str] = self._truncate( lowerCamelCase__ ,max_length=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,truncation=lowerCamelCase__ ,) truncated_inputs.append(lowerCamelCase__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length _UpperCamelCase : Union[str, Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) _UpperCamelCase : Any = PaddingStrategy.MAX_LENGTH _UpperCamelCase : Optional[Any] = {} for i in range(lowerCamelCase__ ): # padding _UpperCamelCase : Any = self._pad( truncated_inputs[i] ,max_length=lowerCamelCase__ ,padding_strategy=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,) for key, value in outputs.items(): if key not in batch_outputs: _UpperCamelCase : Dict = [] if value.dtype is np.dtype(np.floataa ): _UpperCamelCase : Any = value.astype(np.floataa ) batch_outputs[key].append(lowerCamelCase__ ) return BatchFeature(lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' _UpperCamelCase : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: _UpperCamelCase : Optional[Any] = len(lowerCamelCase__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : str = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : str = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowerCamelCase__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: _UpperCamelCase : Tuple = np.ones(len(lowerCamelCase__ ) ,dtype=np.intaa ) if needs_to_be_padded: _UpperCamelCase : Dict = max_length - len(lowerCamelCase__ ) if self.padding_side == "right": if return_attention_mask: _UpperCamelCase : Optional[int] = np.pad( processed_features['attention_mask'] ,(0, difference) ) _UpperCamelCase : Union[str, Any] = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) _UpperCamelCase : List[Any] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: _UpperCamelCase : List[Any] = np.pad( processed_features['attention_mask'] ,(difference, 0) ) _UpperCamelCase : List[Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) _UpperCamelCase : List[str] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) _UpperCamelCase : int = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : Optional[int] = len(lowerCamelCase__ ) > max_length if needs_to_be_truncated: _UpperCamelCase : Dict = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: _UpperCamelCase : Optional[Any] = processed_features['attention_mask'][:max_length] return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : int=False ,lowerCamelCase__ : Optional[Any]=None ): '''simple docstring''' # Get padding strategy if padding is not False: if padding is True: _UpperCamelCase : Optional[Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Tuple = PaddingStrategy(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = padding else: _UpperCamelCase : List[Any] = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
83
1
'''simple docstring''' from statistics import mean, stdev def A__ ( UpperCAmelCase_ , UpperCAmelCase_ = 3 ): _UpperCamelCase : List[Any] = min(UpperCAmelCase_ ) _UpperCamelCase : Dict = max(UpperCAmelCase_ ) # normalize data return [round((x - x_min) / (x_max - x_min) , UpperCAmelCase_ ) for x in data] def A__ ( UpperCAmelCase_ , UpperCAmelCase_ = 3 ): _UpperCamelCase : str = mean(UpperCAmelCase_ ) _UpperCamelCase : Tuple = stdev(UpperCAmelCase_ ) # standardize data return [round((x - mu) / (sigma) , UpperCAmelCase_ ) for x in data]
83
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
1
'''simple docstring''' import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) _UpperCamelCase : Any = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(lowerCamelCase__ ) _UpperCamelCase : str = -1 _UpperCamelCase : Optional[Any] = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = model.generate(lowerCamelCase__ ,max_new_tokens=10 ,do_sample=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: _UpperCamelCase : Any = TextStreamer(lowerCamelCase__ ) model.generate(lowerCamelCase__ ,max_new_tokens=10 ,do_sample=lowerCamelCase__ ,streamer=lowerCamelCase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer _UpperCamelCase : Optional[int] = cs.out[:-1] self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) _UpperCamelCase : Union[str, Any] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(lowerCamelCase__ ) _UpperCamelCase : Optional[int] = -1 _UpperCamelCase : Optional[int] = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(lowerCamelCase__ ) _UpperCamelCase : List[Any] = model.generate(lowerCamelCase__ ,max_new_tokens=10 ,do_sample=lowerCamelCase__ ) _UpperCamelCase : int = tokenizer.decode(greedy_ids[0] ) _UpperCamelCase : List[str] = TextIteratorStreamer(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {'input_ids': input_ids, 'max_new_tokens': 10, 'do_sample': False, 'streamer': streamer} _UpperCamelCase : Optional[Any] = Thread(target=model.generate ,kwargs=lowerCamelCase__ ) thread.start() _UpperCamelCase : Tuple = '' for new_text in streamer: streamer_text += new_text self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) _UpperCamelCase : int = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(lowerCamelCase__ ) _UpperCamelCase : List[Any] = -1 _UpperCamelCase : Dict = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(lowerCamelCase__ ) _UpperCamelCase : str = model.generate(lowerCamelCase__ ,max_new_tokens=10 ,do_sample=lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = greedy_ids[:, input_ids.shape[1] :] _UpperCamelCase : str = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: _UpperCamelCase : Optional[int] = TextStreamer(lowerCamelCase__ ,skip_prompt=lowerCamelCase__ ) model.generate(lowerCamelCase__ ,max_new_tokens=10 ,do_sample=lowerCamelCase__ ,streamer=lowerCamelCase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer _UpperCamelCase : Optional[int] = cs.out[:-1] self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them _UpperCamelCase : Dict = AutoTokenizer.from_pretrained('distilgpt2' ) _UpperCamelCase : int = AutoModelForCausalLM.from_pretrained('distilgpt2' ).to(lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = -1 _UpperCamelCase : List[str] = torch.ones((1, 5) ,device=lowerCamelCase__ ).long() * model.config.bos_token_id with CaptureStdout() as cs: _UpperCamelCase : Dict = TextStreamer(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ ) model.generate(lowerCamelCase__ ,max_new_tokens=1 ,do_sample=lowerCamelCase__ ,streamer=lowerCamelCase__ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token _UpperCamelCase : Dict = cs.out[:-1] # Remove the final "\n" _UpperCamelCase : str = tokenizer(lowerCamelCase__ ,return_tensors='pt' ) self.assertEqual(streamer_text_tokenized.input_ids.shape ,(1, 1) ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) _UpperCamelCase : Optional[Any] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(lowerCamelCase__ ) _UpperCamelCase : Any = -1 _UpperCamelCase : Dict = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(lowerCamelCase__ ) _UpperCamelCase : str = TextIteratorStreamer(lowerCamelCase__ ,timeout=0.0_0_1 ) _UpperCamelCase : Any = {'input_ids': input_ids, 'max_new_tokens': 10, 'do_sample': False, 'streamer': streamer} _UpperCamelCase : Optional[int] = Thread(target=model.generate ,kwargs=lowerCamelCase__ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(lowerCamelCase__ ): _UpperCamelCase : int = '' for new_text in streamer: streamer_text += new_text
83
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowercase__ ( lowercase ): @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Dict = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : Dict = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : str = '1' _UpperCamelCase : Union[str, Any] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : List[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : List[Any] = self.get_env() _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # next emulate no network _UpperCamelCase : Dict = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : Dict = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : int = '\nfrom transformers import pipeline\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _UpperCamelCase : Union[str, Any] = self.get_env() _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Tuple = [sys.executable, '-c', '\n'.join([load, mock, run] )] _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,1 ,result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' ,result.stderr.decode().replace('\n' ,'' ) ,) @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = '\nfrom transformers import AutoModel\n ' _UpperCamelCase : int = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Any = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : Optional[int] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() )
83
1
'''simple docstring''' def A__ ( UpperCAmelCase_=2_8_1_2_3 ): _UpperCamelCase : Any = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i _UpperCamelCase : List[Any] = set() _UpperCamelCase : str = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(UpperCAmelCase_ ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
83
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase__ ( unittest.TestCase ): def __init__( self : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=99 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=512 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : int=0.0_2 ,lowerCamelCase__ : int=4 ,): '''simple docstring''' _UpperCamelCase : List[Any] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Optional[Any] = is_training _UpperCamelCase : Optional[int] = use_attention_mask _UpperCamelCase : Any = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : str = intermediate_size _UpperCamelCase : int = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Optional[int] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_choices def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Union[str, Any] = None if self.use_attention_mask: _UpperCamelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=lowerCamelCase__ ,) return config, input_ids, attention_mask def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Optional[int] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase : Dict = model_class_name.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase : Tuple = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] _UpperCamelCase : Any = (1, 11, 768) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1E-4 ) )
83
1
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class lowercase__ ( lowercase ): @staticmethod @abstractmethod def UpperCamelCase_ ( lowerCamelCase__ : ArgumentParser ): '''simple docstring''' raise NotImplementedError() @abstractmethod def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' raise NotImplementedError()
83
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer snake_case_ : List[Any] = logging.get_logger(__name__) class lowercase__ ( lowercase ): lowercase__ = """AutoTokenizer""" lowercase__ = ["""tokenizer"""] lowercase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self : List[str] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Tuple=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : Dict = speaker_embeddings @classmethod def UpperCamelCase_ ( cls : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : str="speaker_embeddings_path.json" ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: _UpperCamelCase : Optional[Any] = get_file_from_repo( lowerCamelCase__ ,lowerCamelCase__ ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(lowerCamelCase__ ,lowerCamelCase__ )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) _UpperCamelCase : Union[str, Any] = None else: with open(lowerCamelCase__ ) as speaker_embeddings_json: _UpperCamelCase : Optional[int] = json.load(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = None _UpperCamelCase : Tuple = AutoTokenizer.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) return cls(tokenizer=lowerCamelCase__ ,speaker_embeddings=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int="speaker_embeddings_path.json" ,lowerCamelCase__ : Dict="speaker_embeddings" ,lowerCamelCase__ : bool = False ,**lowerCamelCase__ : Tuple ,): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ,'v2' ) ,exist_ok=lowerCamelCase__ ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Optional[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCamelCase : Any = self._load_voice_preset(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] ,lowerCamelCase__ ,F'{prompt_key}_{key}' ) ,voice_preset[key] ,allow_pickle=lowerCamelCase__ ,) _UpperCamelCase : List[str] = os.path.join(lowerCamelCase__ ,F'{prompt_key}_{key}.npy' ) _UpperCamelCase : str = tmp_dict with open(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ) ,'w' ) as fp: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) super().save_pretrained(lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : str = None ,**lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.speaker_embeddings[voice_preset] _UpperCamelCase : Union[str, Any] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) _UpperCamelCase : Dict = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' ,'/' ) ,voice_preset_paths[key] ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" ,"/" ) ,voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) _UpperCamelCase : List[str] = np.load(lowerCamelCase__ ) return voice_preset_dict def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] ,np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self : Any ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any="pt" ,lowerCamelCase__ : Dict=256 ,lowerCamelCase__ : int=False ,lowerCamelCase__ : int=True ,lowerCamelCase__ : List[str]=False ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' if voice_preset is not None and not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if ( isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCamelCase : Optional[int] = self._load_voice_preset(lowerCamelCase__ ) else: if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and not voice_preset.endswith('.npz' ): _UpperCamelCase : Tuple = voice_preset + '.npz' _UpperCamelCase : str = np.load(lowerCamelCase__ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.tokenizer( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,padding='max_length' ,max_length=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,**lowerCamelCase__ ,) if voice_preset is not None: _UpperCamelCase : Optional[Any] = voice_preset return encoded_text
83
1
'''simple docstring''' import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() snake_case_ : Optional[int] = logging.get_logger(__name__) def A__ ( UpperCAmelCase_ ): print('Loading config file...' ) def flatten_yaml_as_dict(UpperCAmelCase_ , UpperCAmelCase_="" , UpperCAmelCase_="." ): _UpperCamelCase : int = [] for k, v in d.items(): _UpperCamelCase : Optional[int] = parent_key + sep + k if parent_key else k if isinstance(UpperCAmelCase_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(UpperCAmelCase_ , UpperCAmelCase_ , sep=UpperCAmelCase_ ).items() ) else: items.append((new_key, v) ) return dict(UpperCAmelCase_ ) _UpperCamelCase : Tuple = argparse.Namespace() with open(UpperCAmelCase_ , 'r' ) as yaml_file: try: _UpperCamelCase : int = yaml.load(UpperCAmelCase_ , Loader=yaml.FullLoader ) _UpperCamelCase : Optional[int] = flatten_yaml_as_dict(UpperCAmelCase_ ) for k, v in flat_cfg.items(): setattr(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) except yaml.YAMLError as exc: logger.error('Error while loading config file: {}. Error message: {}'.format(UpperCAmelCase_ , str(UpperCAmelCase_ ) ) ) return config def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Dict = MobileViTVaConfig() _UpperCamelCase : Union[str, Any] = False # dataset if task_name.startswith('imagenet1k_' ): _UpperCamelCase : Dict = 1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: _UpperCamelCase : List[str] = 3_8_4 else: _UpperCamelCase : Union[str, Any] = 2_5_6 _UpperCamelCase : Union[str, Any] = 'imagenet-1k-id2label.json' elif task_name.startswith('imagenet21k_to_1k_' ): _UpperCamelCase : Optional[int] = 2_1_0_0_0 if int(task_name.strip().split('_' )[-1] ) == 3_8_4: _UpperCamelCase : Tuple = 3_8_4 else: _UpperCamelCase : Dict = 2_5_6 _UpperCamelCase : Any = 'imagenet-22k-id2label.json' elif task_name.startswith('ade20k_' ): _UpperCamelCase : Optional[Any] = 1_5_1 _UpperCamelCase : Any = 5_1_2 _UpperCamelCase : Optional[Any] = 'ade20k-id2label.json' _UpperCamelCase : Optional[int] = True elif task_name.startswith('voc_' ): _UpperCamelCase : Any = 2_1 _UpperCamelCase : List[str] = 5_1_2 _UpperCamelCase : Union[str, Any] = 'pascal-voc-id2label.json' _UpperCamelCase : Optional[Any] = True # orig_config _UpperCamelCase : Dict = load_orig_config_file(UpperCAmelCase_ ) assert getattr(UpperCAmelCase_ , 'model.classification.name' , -1 ) == "mobilevit_v2", "Invalid model" _UpperCamelCase : Dict = getattr(UpperCAmelCase_ , 'model.classification.mitv2.width_multiplier' , 1.0 ) assert ( getattr(UpperCAmelCase_ , 'model.classification.mitv2.attn_norm_layer' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" _UpperCamelCase : Tuple = getattr(UpperCAmelCase_ , 'model.classification.activation.name' , 'swish' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: _UpperCamelCase : List[str] = getattr(UpperCAmelCase_ , 'model.segmentation.output_stride' , 1_6 ) if "_deeplabv3" in task_name: _UpperCamelCase : Optional[Any] = getattr(UpperCAmelCase_ , 'model.segmentation.deeplabv3.aspp_rates' , [1_2, 2_4, 3_6] ) _UpperCamelCase : str = getattr(UpperCAmelCase_ , 'model.segmentation.deeplabv3.aspp_out_channels' , 5_1_2 ) _UpperCamelCase : List[str] = getattr(UpperCAmelCase_ , 'model.segmentation.deeplabv3.aspp_dropout' , 0.1 ) # id2label _UpperCamelCase : Dict = 'huggingface/label-files' _UpperCamelCase : Any = json.load(open(hf_hub_download(UpperCAmelCase_ , UpperCAmelCase_ , repo_type='dataset' ) , 'r' ) ) _UpperCamelCase : Tuple = {int(UpperCAmelCase_ ): v for k, v in idalabel.items()} _UpperCamelCase : Optional[Any] = idalabel _UpperCamelCase : int = {v: k for k, v in idalabel.items()} return config def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : List[Any] = dct.pop(UpperCAmelCase_ ) _UpperCamelCase : Union[str, Any] = val def A__ ( UpperCAmelCase_ , UpperCAmelCase_=False ): if base_model: _UpperCamelCase : Optional[int] = '' else: _UpperCamelCase : Union[str, Any] = 'mobilevitv2.' _UpperCamelCase : int = [] for k in state_dict.keys(): if k[:8] == "encoder.": _UpperCamelCase : int = k[8:] else: _UpperCamelCase : Any = k if ".block." in k: _UpperCamelCase : Optional[int] = k_new.replace('.block.' , '.' ) if ".conv." in k: _UpperCamelCase : Union[str, Any] = k_new.replace('.conv.' , '.convolution.' ) if ".norm." in k: _UpperCamelCase : List[Any] = k_new.replace('.norm.' , '.normalization.' ) if "conv_1." in k: _UpperCamelCase : Union[str, Any] = k_new.replace('conv_1.' , f'{model_prefix}conv_stem.' ) for i in [1, 2]: if f'layer_{i}.' in k: _UpperCamelCase : Optional[int] = k_new.replace(f'layer_{i}.' , f'{model_prefix}encoder.layer.{i-1}.layer.' ) if ".exp_1x1." in k: _UpperCamelCase : Dict = k_new.replace('.exp_1x1.' , '.expand_1x1.' ) if ".red_1x1." in k: _UpperCamelCase : List[str] = k_new.replace('.red_1x1.' , '.reduce_1x1.' ) for i in [3, 4, 5]: if f'layer_{i}.0.' in k: _UpperCamelCase : Tuple = k_new.replace(f'layer_{i}.0.' , f'{model_prefix}encoder.layer.{i-1}.downsampling_layer.' ) if f'layer_{i}.1.local_rep.0.' in k: _UpperCamelCase : str = k_new.replace(f'layer_{i}.1.local_rep.0.' , f'{model_prefix}encoder.layer.{i-1}.conv_kxk.' ) if f'layer_{i}.1.local_rep.1.' in k: _UpperCamelCase : Optional[Any] = k_new.replace(f'layer_{i}.1.local_rep.1.' , f'{model_prefix}encoder.layer.{i-1}.conv_1x1.' ) for i in [3, 4, 5]: if i == 3: _UpperCamelCase : List[Any] = [0, 1] elif i == 4: _UpperCamelCase : Any = [0, 1, 2, 3] elif i == 5: _UpperCamelCase : List[Any] = [0, 1, 2] for j in j_in: if f'layer_{i}.1.global_rep.{j}.' in k: _UpperCamelCase : Tuple = k_new.replace( f'layer_{i}.1.global_rep.{j}.' , f'{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.' ) if f'layer_{i}.1.global_rep.{j+1}.' in k: _UpperCamelCase : Tuple = k_new.replace( f'layer_{i}.1.global_rep.{j+1}.' , f'{model_prefix}encoder.layer.{i-1}.layernorm.' ) if f'layer_{i}.1.conv_proj.' in k: _UpperCamelCase : List[str] = k_new.replace(f'layer_{i}.1.conv_proj.' , f'{model_prefix}encoder.layer.{i-1}.conv_projection.' ) if "pre_norm_attn.0." in k: _UpperCamelCase : str = k_new.replace('pre_norm_attn.0.' , 'layernorm_before.' ) if "pre_norm_attn.1." in k: _UpperCamelCase : Optional[int] = k_new.replace('pre_norm_attn.1.' , 'attention.' ) if "pre_norm_ffn.0." in k: _UpperCamelCase : Tuple = k_new.replace('pre_norm_ffn.0.' , 'layernorm_after.' ) if "pre_norm_ffn.1." in k: _UpperCamelCase : Any = k_new.replace('pre_norm_ffn.1.' , 'ffn.conv1.' ) if "pre_norm_ffn.3." in k: _UpperCamelCase : Any = k_new.replace('pre_norm_ffn.3.' , 'ffn.conv2.' ) if "classifier.1." in k: _UpperCamelCase : Optional[int] = k_new.replace('classifier.1.' , 'classifier.' ) if "seg_head." in k: _UpperCamelCase : str = k_new.replace('seg_head.' , 'segmentation_head.' ) if ".aspp_layer." in k: _UpperCamelCase : Dict = k_new.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in k: _UpperCamelCase : Union[str, Any] = k_new.replace('.aspp_pool.' , '.' ) rename_keys.append((k, k_new) ) return rename_keys def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = [] for k in state_dict.keys(): if k.startswith('seg_head.aux_head.' ): keys_to_ignore.append(UpperCAmelCase_ ) for k in keys_to_ignore: state_dict.pop(UpperCAmelCase_ , UpperCAmelCase_ ) def A__ ( ): _UpperCamelCase : Any = 'http://images.cocodataset.org/val2017/000000039769.jpg' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" _UpperCamelCase : str = Image.open(requests.get(UpperCAmelCase_ , stream=UpperCAmelCase_ ).raw ) return im @torch.no_grad() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Union[str, Any] = get_mobilevitva_config(UpperCAmelCase_ , UpperCAmelCase_ ) # load original state_dict _UpperCamelCase : Union[str, Any] = torch.load(UpperCAmelCase_ , map_location='cpu' ) # load huggingface model if task_name.startswith('ade20k_' ) or task_name.startswith('voc_' ): _UpperCamelCase : Optional[int] = MobileViTVaForSemanticSegmentation(UpperCAmelCase_ ).eval() _UpperCamelCase : Optional[int] = False else: _UpperCamelCase : Union[str, Any] = MobileViTVaForImageClassification(UpperCAmelCase_ ).eval() _UpperCamelCase : str = False # remove and rename some keys of load the original model _UpperCamelCase : Tuple = checkpoint remove_unused_keys(UpperCAmelCase_ ) _UpperCamelCase : Tuple = create_rename_keys(UpperCAmelCase_ , base_model=UpperCAmelCase_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # load modified state_dict model.load_state_dict(UpperCAmelCase_ ) # Check outputs on an image, prepared by MobileViTImageProcessor _UpperCamelCase : Union[str, Any] = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 3_2 ) _UpperCamelCase : Any = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCamelCase : Optional[Any] = model(**UpperCAmelCase_ ) # verify classification model if task_name.startswith('imagenet' ): _UpperCamelCase : List[str] = outputs.logits _UpperCamelCase : List[Any] = logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('imagenet1k_256' ) and config.width_multiplier == 1.0: # expected_logits for base variant _UpperCamelCase : List[str] = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ) assert torch.allclose(logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) Path(UpperCAmelCase_ ).mkdir(exist_ok=UpperCAmelCase_ ) print(f'Saving model {task_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase_ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": snake_case_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='imagenet1k_256', type=str, help=( 'Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . ' '\n Classification (ImageNet-1k)\n - MobileViTV2 (256x256) : imagenet1k_256\n - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384\n - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) :\n imagenet21k_to_1k_256\n - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on\n ImageNet-1k 384x384) : imagenet21k_to_1k_384\n Segmentation\n - ADE20K Dataset : ade20k_deeplabv3\n - Pascal VOC 2012 Dataset: voc_deeplabv3\n ' ), choices=[ 'imagenet1k_256', 'imagenet1k_384', 'imagenet21k_to_1k_256', 'imagenet21k_to_1k_384', 'ade20k_deeplabv3', 'voc_deeplabv3', ], ) parser.add_argument( '--orig_checkpoint_path', required=True, type=str, help='Path to the original state dict (.pt file).' ) parser.add_argument('--orig_config_path', required=True, type=str, help='Path to the original config file.') parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) snake_case_ : int = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
83
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin snake_case_ : Tuple = random.Random() def A__ ( UpperCAmelCase_ , UpperCAmelCase_=1.0 , UpperCAmelCase_=None , UpperCAmelCase_=None ): if rng is None: _UpperCamelCase : Dict = global_rng _UpperCamelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int=7 ,lowerCamelCase__ : str=400 ,lowerCamelCase__ : int=2000 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=16000 ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Optional[int]=True ,): '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Union[str, Any] = batch_size _UpperCamelCase : List[str] = min_seq_length _UpperCamelCase : Optional[int] = max_seq_length _UpperCamelCase : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _UpperCamelCase : List[str] = feature_size _UpperCamelCase : List[str] = padding_value _UpperCamelCase : List[Any] = sampling_rate _UpperCamelCase : Dict = return_attention_mask _UpperCamelCase : Tuple = do_normalize def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : Tuple=False ): '''simple docstring''' def _flatten(lowerCamelCase__ : Optional[Any] ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _UpperCamelCase : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _UpperCamelCase : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: _UpperCamelCase : int = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = WavaVecaFeatureExtractor def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): '''simple docstring''' self.assertTrue(np.all(np.mean(lowerCamelCase__ ,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ,axis=0 ) - 1 ) < 1E-3 ) ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' # Tests that all call wrap to encode_plus and batch_encode_plus _UpperCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _UpperCamelCase : int = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Tuple = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test not batched input _UpperCamelCase : Tuple = feat_extract(speech_inputs[0] ,return_tensors='np' ).input_values _UpperCamelCase : Any = feat_extract(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test batched _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : Optional[int] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _UpperCamelCase : str = [floats_list((1, x) )[0] for x in (800, 800, 800)] _UpperCamelCase : str = np.asarray(lowerCamelCase__ ) _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : int = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : str = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : List[str] = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,return_tensors='np' ) _UpperCamelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[str] = range(800 ,1400 ,200 ) _UpperCamelCase : List[str] = [floats_list((1, x) )[0] for x in lengths] _UpperCamelCase : Optional[Any] = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : str = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,max_length=lowerCamelCase__ ,padding=lowerCamelCase__ ) _UpperCamelCase : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Union[str, Any] = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='max_length' ,return_tensors='np' ) _UpperCamelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : int = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Any = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=2000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase_ ( self : Any ): '''simple docstring''' import torch _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = np.random.rand(100 ).astype(np.floataa ) _UpperCamelCase : Dict = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _UpperCamelCase : Optional[int] = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _UpperCamelCase : Tuple = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: _UpperCamelCase : Optional[int] = WavaVecaConfig.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase__ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask ,config.feat_extract_norm == 'layer' )
83
1
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1 , UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : int = 1 _UpperCamelCase : Union[str, Any] = 0 for divide_by_number in range(UpperCAmelCase_ , digit + 1 ): _UpperCamelCase : list[int] = [] _UpperCamelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = len(UpperCAmelCase_ ) _UpperCamelCase : List[Any] = divide_by_number else: has_been_divided.append(UpperCAmelCase_ ) _UpperCamelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
83
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case_ : str = logging.get_logger(__name__) snake_case_ : Any = {'vocab_file': 'spiece.model'} snake_case_ : List[str] = { 'vocab_file': { 'bert_for_seq_generation': ( 'https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model' ), } } snake_case_ : str = {'bert_for_seq_generation': 512} class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = [] lowercase__ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[int] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[str]="<s>" ,lowerCamelCase__ : List[str]="</s>" ,lowerCamelCase__ : int="<unk>" ,lowerCamelCase__ : Dict="<pad>" ,lowerCamelCase__ : Optional[Any]="<::::>" ,lowerCamelCase__ : Optional[Dict[str, Any]] = None ,**lowerCamelCase__ : List[str] ,): '''simple docstring''' _UpperCamelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,sp_model_kwargs=self.sp_model_kwargs ,**lowerCamelCase__ ,) _UpperCamelCase : Optional[Any] = vocab_file _UpperCamelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase__ ) @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' return self.sp_model.get_piece_size() def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : Optional[int] = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ): '''simple docstring''' _UpperCamelCase : Optional[Any] = self.__dict__.copy() _UpperCamelCase : Optional[Any] = None return state def __setstate__( self : List[Any] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' _UpperCamelCase : Optional[int] = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _UpperCamelCase : int = {} _UpperCamelCase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : str ): '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ ,out_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : str ,lowerCamelCase__ : List[str] ): '''simple docstring''' return self.sp_model.piece_to_id(lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : Any = self.sp_model.IdToPiece(lowerCamelCase__ ) return token def UpperCamelCase_ ( self : int ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = [] _UpperCamelCase : Dict = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCamelCase__ ) + token _UpperCamelCase : int = [] else: current_sub_tokens.append(lowerCamelCase__ ) out_string += self.sp_model.decode(lowerCamelCase__ ) return out_string.strip() def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ ,'wb' ) as fi: _UpperCamelCase : int = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
83
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if num < 0: return False _UpperCamelCase : int = num _UpperCamelCase : int = 0 while num > 0: _UpperCamelCase : str = rev_num * 1_0 + (num % 1_0) num //= 1_0 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
83
1
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = abs(UpperCAmelCase_ ) _UpperCamelCase : int = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = abs(UpperCAmelCase_ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def A__ ( UpperCAmelCase_ ): return sum(int(UpperCAmelCase_ ) for c in str(abs(UpperCAmelCase_ ) ) ) def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) -> None: _UpperCamelCase : str = f'{func.__name__}({value})' _UpperCamelCase : Tuple = timeit(f'__main__.{call}' , setup='import __main__' ) print(f'{call:56} = {func(UpperCAmelCase_ )} -- {timing:.4f} seconds' ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
83
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) snake_case_ : Dict = { 'configuration_clip': [ 'CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CLIPConfig', 'CLIPOnnxConfig', 'CLIPTextConfig', 'CLIPVisionConfig', ], 'processing_clip': ['CLIPProcessor'], 'tokenization_clip': ['CLIPTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : List[Any] = ['CLIPTokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : int = ['CLIPFeatureExtractor'] snake_case_ : Optional[int] = ['CLIPImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : str = [ 'CLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'CLIPModel', 'CLIPPreTrainedModel', 'CLIPTextModel', 'CLIPTextModelWithProjection', 'CLIPVisionModel', 'CLIPVisionModelWithProjection', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Any = [ 'TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFCLIPModel', 'TFCLIPPreTrainedModel', 'TFCLIPTextModel', 'TFCLIPVisionModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = [ 'FlaxCLIPModel', 'FlaxCLIPPreTrainedModel', 'FlaxCLIPTextModel', 'FlaxCLIPTextPreTrainedModel', 'FlaxCLIPVisionModel', 'FlaxCLIPVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys snake_case_ : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
1
'''simple docstring''' import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowercase__ ( lowercase , unittest.TestCase ): # TODO: is there an appropriate internal test set? lowercase__ = """ssube/stable-diffusion-x4-upscaler-onnx""" def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : Optional[int]=0 ): '''simple docstring''' _UpperCamelCase : Any = floats_tensor((1, 3, 128, 128) ,rng=random.Random(lowerCamelCase__ ) ) _UpperCamelCase : List[str] = torch.manual_seed(lowerCamelCase__ ) _UpperCamelCase : int = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Optional[int] = self.get_dummy_inputs() _UpperCamelCase : str = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[Any] = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Dict = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Optional[int] = PNDMScheduler.from_config(pipe.scheduler.config ,skip_prk_steps=lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : int = self.get_dummy_inputs() _UpperCamelCase : Tuple = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Optional[Any] = np.array( [0.6_8_9_8_8_9_2, 0.5_9_2_4_0_5_5_6, 0.5_2_4_9_9_5_2_7, 0.5_8_8_6_6_2_1_5, 0.5_2_2_5_8_2_3_5, 0.5_2_5_7_2_7_1_5, 0.6_2_4_1_4_4_7_3, 0.6_1_7_4_3_8_7, 0.6_2_1_4_9_6_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = self.get_dummy_inputs() _UpperCamelCase : Union[str, Any] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Dict = np.array( [0.7_6_5_9_2_7_8, 0.7_6_4_3_7_6_6_4, 0.7_5_5_7_9_1_0_7, 0.7_6_9_1_1_1_6, 0.7_7_6_6_6_9_8_6, 0.7_7_2_7_6_7_2, 0.7_7_5_8_6_6_4, 0.7_8_1_2_2_2_6, 0.7_6_9_4_2_5_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Optional[Any] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Tuple = self.get_dummy_inputs() _UpperCamelCase : Optional[int] = pipe(**lowerCamelCase__ ).images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Union[str, Any] = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : str = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint ,provider='CPUExecutionProvider' ) _UpperCamelCase : Tuple = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Dict = self.get_dummy_inputs() _UpperCamelCase : int = pipe(**lowerCamelCase__ ).images _UpperCamelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : List[str] = np.array( [0.7_7_4_2_4_4_9_6, 0.7_7_3_6_0_1, 0.7_6_4_5_2_8_8, 0.7_7_6_9_5_9_8, 0.7_7_7_2_7_3_9, 0.7_7_3_8_6_8_8, 0.7_8_1_8_7_2_3_3, 0.7_7_8_7_9_5_8_4, 0.7_6_7_0_4_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class lowercase__ ( unittest.TestCase ): @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = ort.SessionOptions() _UpperCamelCase : Tuple = False return options def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) _UpperCamelCase : Union[str, Any] = init_image.resize((128, 128) ) # using the PNDM scheduler by default _UpperCamelCase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Dict = 'A fantasy landscape, trending on artstation' _UpperCamelCase : Optional[int] = torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipe( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=10 ,generator=lowerCamelCase__ ,output_type='np' ,) _UpperCamelCase : Union[str, Any] = output.images _UpperCamelCase : Any = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCamelCase : List[Any] = np.array([0.4_8_8_3, 0.4_9_4_7, 0.4_9_8_0, 0.4_9_7_5, 0.4_9_8_2, 0.4_9_8_0, 0.5_0_0_0, 0.5_0_0_6, 0.4_9_7_2] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) _UpperCamelCase : List[Any] = init_image.resize((128, 128) ) _UpperCamelCase : Optional[Any] = LMSDiscreteScheduler.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' ,subfolder='scheduler' ) _UpperCamelCase : Any = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' ,scheduler=lowerCamelCase__ ,provider=self.gpu_provider ,sess_options=self.gpu_options ,) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Optional[int] = 'A fantasy landscape, trending on artstation' _UpperCamelCase : List[str] = torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = pipe( prompt=lowerCamelCase__ ,image=lowerCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=20 ,generator=lowerCamelCase__ ,output_type='np' ,) _UpperCamelCase : List[Any] = output.images _UpperCamelCase : List[str] = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCamelCase : List[str] = np.array( [0.5_0_1_7_3_7_5_3, 0.5_0_2_2_3_3_5_6, 0.5_0_2_0_3_9, 0.5_0_2_3_3_0_3_6, 0.5_0_2_3_7_2_5, 0.5_0_2_2_6_0_1, 0.5_0_1_8_7_5_8, 0.5_0_2_3_4_0_8_5, 0.5_0_2_4_1_5_6_6] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
83
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
1
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase__ ( lowercase ): lowercase__ = ["""image_processor""", """tokenizer"""] lowercase__ = """ViTImageProcessor""" lowercase__ = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : Optional[int] ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Optional[Any]=None ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' _UpperCamelCase : int = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' ,lowerCamelCase__ ,) _UpperCamelCase : List[Any] = kwargs.pop('feature_extractor' ) _UpperCamelCase : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(lowerCamelCase__ ,lowerCamelCase__ ) def __call__( self : List[Any] ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Any=None ,lowerCamelCase__ : str=None ,lowerCamelCase__ : List[Any]=None ,**lowerCamelCase__ : List[str] ): '''simple docstring''' if text is None and visual_prompt is None and images is None: raise ValueError('You have to specify either text, visual prompt or images.' ) if text is not None and visual_prompt is not None: raise ValueError('You have to specify exactly one type of prompt. Either text or visual prompt.' ) if text is not None: _UpperCamelCase : Optional[int] = self.tokenizer(lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) if visual_prompt is not None: _UpperCamelCase : Any = self.image_processor(lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) if images is not None: _UpperCamelCase : Union[str, Any] = self.image_processor(lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,**lowerCamelCase__ ) if visual_prompt is not None and images is not None: _UpperCamelCase : Dict = { 'pixel_values': image_features.pixel_values, 'conditional_pixel_values': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: _UpperCamelCase : Optional[int] = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: _UpperCamelCase : Dict = { 'conditional_pixel_values': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**lowerCamelCase__ ) ,tensor_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,*lowerCamelCase__ : Dict ,**lowerCamelCase__ : int ): '''simple docstring''' return self.tokenizer.batch_decode(*lowerCamelCase__ ,**lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,*lowerCamelCase__ : Any ,**lowerCamelCase__ : int ): '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase__ ,**lowerCamelCase__ ) @property def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' ,lowerCamelCase__ ,) return self.image_processor_class @property def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' ,lowerCamelCase__ ,) return self.image_processor
83
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
1
'''simple docstring''' import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset snake_case_ : Union[str, Any] = 'bert-base-cased' snake_case_ : Any = 'google/pegasus-xsum' snake_case_ : Tuple = [' Sam ate lunch today.', 'Sams lunch ingredients.'] snake_case_ : List[str] = ['A very interesting story about what I ate for lunch.', 'Avocado, celery, turkey, coffee'] snake_case_ : int = 'patrickvonplaten/t5-tiny-random' snake_case_ : Union[str, Any] = 'sshleifer/bart-tiny-random' snake_case_ : Any = 'sshleifer/tiny-mbart' snake_case_ : Any = 'sshleifer/tiny-marian-en-de' def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Any = '\n'.join(UpperCAmelCase_ ) Path(UpperCAmelCase_ ).open('w' ).writelines(UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): for split in ["train", "val", "test"]: _dump_articles(os.path.join(UpperCAmelCase_ , f'{split}.source' ) , UpperCAmelCase_ ) _dump_articles(os.path.join(UpperCAmelCase_ , f'{split}.target' ) , UpperCAmelCase_ ) return tmp_dir class lowercase__ ( lowercase ): @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] ,) @slow def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Tuple = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : int = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) _UpperCamelCase : Optional[Any] = max(len(tokenizer.encode(lowerCamelCase__ ) ) for a in ARTICLES ) _UpperCamelCase : str = max(len(tokenizer.encode(lowerCamelCase__ ) ) for a in SUMMARIES ) _UpperCamelCase : Dict = 4 _UpperCamelCase : List[Any] = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated _UpperCamelCase , _UpperCamelCase : int = 'ro_RO', 'de_DE' # ignored for all but mbart, but never causes error. _UpperCamelCase : Optional[Any] = SeqaSeqDataset( lowerCamelCase__ ,data_dir=lowerCamelCase__ ,type_path='train' ,max_source_length=lowerCamelCase__ ,max_target_length=lowerCamelCase__ ,src_lang=lowerCamelCase__ ,tgt_lang=lowerCamelCase__ ,) _UpperCamelCase : Union[str, Any] = DataLoader(lowerCamelCase__ ,batch_size=2 ,collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(lowerCamelCase__ ,lowerCamelCase__ ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place _UpperCamelCase : Tuple = shift_tokens_right(batch['labels'] ,tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Any ): '''simple docstring''' _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) _UpperCamelCase : Optional[Any] = max(len(tokenizer.encode(lowerCamelCase__ ) ) for a in ARTICLES ) _UpperCamelCase : Dict = max(len(tokenizer.encode(lowerCamelCase__ ) ) for a in SUMMARIES ) _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[str] = LegacySeqaSeqDataset( lowerCamelCase__ ,data_dir=lowerCamelCase__ ,type_path='train' ,max_source_length=20 ,max_target_length=lowerCamelCase__ ,) _UpperCamelCase : Any = DataLoader(lowerCamelCase__ ,batch_size=2 ,collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 20 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : str = AutoTokenizer.from_pretrained('facebook/mbart-large-cc25' ) _UpperCamelCase : Dict = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) _UpperCamelCase : Dict = tmp_dir.joinpath('train.source' ).open().readlines() _UpperCamelCase : Any = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(lowerCamelCase__ ,lowerCamelCase__ ,128 ,lowerCamelCase__ ) _UpperCamelCase : Tuple = {x.name for x in tmp_dir.iterdir()} _UpperCamelCase : Dict = {x.name for x in save_dir.iterdir()} _UpperCamelCase : Union[str, Any] = save_dir.joinpath('train.source' ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(lowerCamelCase__ ) < len(lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1 assert len(packed_examples[0] ) == sum(len(lowerCamelCase__ ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE ,reason='This test requires fairseq' ) def UpperCamelCase_ ( self : str ): '''simple docstring''' if not FAIRSEQ_AVAILABLE: return _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Dict = self._get_dataset(max_len=64 ) _UpperCamelCase : Optional[Any] = 64 _UpperCamelCase : str = ds.make_dynamic_sampler(lowerCamelCase__ ,required_batch_size_multiple=lowerCamelCase__ ) _UpperCamelCase : Tuple = [len(lowerCamelCase__ ) for x in batch_sampler] assert len(set(lowerCamelCase__ ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(lowerCamelCase__ ) == len(lowerCamelCase__ ) # no dropped or added examples _UpperCamelCase : str = DataLoader(lowerCamelCase__ ,batch_sampler=lowerCamelCase__ ,collate_fn=ds.collate_fn ,num_workers=2 ) _UpperCamelCase : Any = [] _UpperCamelCase : Tuple = [] for batch in data_loader: _UpperCamelCase : str = batch['input_ids'].shape _UpperCamelCase : List[Any] = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple _UpperCamelCase : Dict = np.product(batch['input_ids'].shape ) num_src_per_batch.append(lowerCamelCase__ ) if num_src_tokens > (max_tokens * 1.1): failures.append(lowerCamelCase__ ) assert num_src_per_batch[0] == max(lowerCamelCase__ ) if failures: raise AssertionError(F'too many tokens in {len(lowerCamelCase__ )} batches' ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Optional[int] = self._get_dataset(max_len=512 ) _UpperCamelCase : Dict = 2 _UpperCamelCase : Any = ds.make_sortish_sampler(lowerCamelCase__ ,shuffle=lowerCamelCase__ ) _UpperCamelCase : List[Any] = DataLoader(lowerCamelCase__ ,batch_size=lowerCamelCase__ ,collate_fn=ds.collate_fn ,num_workers=2 ) _UpperCamelCase : List[str] = DataLoader(lowerCamelCase__ ,batch_size=lowerCamelCase__ ,collate_fn=ds.collate_fn ,num_workers=2 ,sampler=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = tokenizer.pad_token_id def count_pad_tokens(lowerCamelCase__ : str ,lowerCamelCase__ : Union[str, Any]="input_ids" ): return [batch[k].eq(lowerCamelCase__ ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(lowerCamelCase__ ,k='labels' ) ) < sum(count_pad_tokens(lowerCamelCase__ ,k='labels' ) ) assert sum(count_pad_tokens(lowerCamelCase__ ) ) < sum(count_pad_tokens(lowerCamelCase__ ) ) assert len(lowerCamelCase__ ) == len(lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[int] ,lowerCamelCase__ : Optional[int]=1000 ,lowerCamelCase__ : Union[str, Any]=128 ): '''simple docstring''' if os.getenv('USE_REAL_DATA' ,lowerCamelCase__ ): _UpperCamelCase : List[Any] = 'examples/seq2seq/wmt_en_ro' _UpperCamelCase : int = max_len * 2 * 64 if not Path(lowerCamelCase__ ).joinpath('train.len' ).exists(): save_len_file(lowerCamelCase__ ,lowerCamelCase__ ) else: _UpperCamelCase : Any = 'examples/seq2seq/test_data/wmt_en_ro' _UpperCamelCase : Any = max_len * 4 save_len_file(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : str = SeqaSeqDataset( lowerCamelCase__ ,data_dir=lowerCamelCase__ ,type_path='train' ,max_source_length=lowerCamelCase__ ,max_target_length=lowerCamelCase__ ,n_obs=lowerCamelCase__ ,) return ds, max_tokens, tokenizer def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[str] = self._get_dataset() _UpperCamelCase : Any = set(DistributedSortishSampler(lowerCamelCase__ ,256 ,num_replicas=2 ,rank=0 ,add_extra_examples=lowerCamelCase__ ) ) _UpperCamelCase : List[Any] = set(DistributedSortishSampler(lowerCamelCase__ ,256 ,num_replicas=2 ,rank=1 ,add_extra_examples=lowerCamelCase__ ) ) assert idsa.intersection(lowerCamelCase__ ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] ,) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(lowerCamelCase__ ,use_fast=lowerCamelCase__ ) if tok_name == MBART_TINY: _UpperCamelCase : Union[str, Any] = SeqaSeqDataset( lowerCamelCase__ ,data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ,type_path='train' ,max_source_length=4 ,max_target_length=8 ,src_lang='EN' ,tgt_lang='FR' ,) _UpperCamelCase : Tuple = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: _UpperCamelCase : Tuple = SeqaSeqDataset( lowerCamelCase__ ,data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ,type_path='train' ,max_source_length=4 ,max_target_length=8 ,) _UpperCamelCase : Any = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(lowerCamelCase__ ) == 1 if tok_name == BART_TINY else len(lowerCamelCase__ ) == 0
83
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
1
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType snake_case_ : List[Any] = logging.get_logger(__name__) snake_case_ : Optional[int] = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """deberta-v2""" def __init__( self : Union[str, Any] ,lowerCamelCase__ : Optional[Any]=128100 ,lowerCamelCase__ : Optional[int]=1536 ,lowerCamelCase__ : Any=24 ,lowerCamelCase__ : Dict=24 ,lowerCamelCase__ : Optional[int]=6144 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[str]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[int]=512 ,lowerCamelCase__ : Dict=0 ,lowerCamelCase__ : Dict=0.0_2 ,lowerCamelCase__ : List[Any]=1E-7 ,lowerCamelCase__ : List[Any]=False ,lowerCamelCase__ : List[str]=-1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : Union[str, Any]=0 ,lowerCamelCase__ : Any="gelu" ,**lowerCamelCase__ : str ,): '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCamelCase : List[str] = hidden_size _UpperCamelCase : str = num_hidden_layers _UpperCamelCase : Union[str, Any] = num_attention_heads _UpperCamelCase : Any = intermediate_size _UpperCamelCase : List[str] = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : List[Any] = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Dict = type_vocab_size _UpperCamelCase : Any = initializer_range _UpperCamelCase : Union[str, Any] = relative_attention _UpperCamelCase : List[str] = max_relative_positions _UpperCamelCase : Dict = pad_token_id _UpperCamelCase : Optional[Any] = position_biased_input # Backwards compatibility if type(lowerCamelCase__ ) == str: _UpperCamelCase : Optional[Any] = [x.strip() for x in pos_att_type.lower().split('|' )] _UpperCamelCase : Any = pos_att_type _UpperCamelCase : int = vocab_size _UpperCamelCase : List[str] = layer_norm_eps _UpperCamelCase : Tuple = kwargs.get('pooler_hidden_size' ,lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = pooler_dropout _UpperCamelCase : str = pooler_hidden_act class lowercase__ ( lowercase ): @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' if self.task == "multiple-choice": _UpperCamelCase : List[str] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _UpperCamelCase : List[Any] = {0: 'batch', 1: 'sequence'} if self._config.type_vocab_size > 0: return OrderedDict( [('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis)] ) else: return OrderedDict([('input_ids', dynamic_axis), ('attention_mask', dynamic_axis)] ) @property def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' return 12 def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] ,lowerCamelCase__ : int = -1 ,lowerCamelCase__ : int = -1 ,lowerCamelCase__ : int = -1 ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional["TensorType"] = None ,lowerCamelCase__ : int = 3 ,lowerCamelCase__ : int = 40 ,lowerCamelCase__ : int = 40 ,lowerCamelCase__ : "PreTrainedTokenizerBase" = None ,): '''simple docstring''' _UpperCamelCase : List[Any] = super().generate_dummy_inputs(preprocessor=lowerCamelCase__ ,framework=lowerCamelCase__ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
83
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
1
'''simple docstring''' import math import sys def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Union[str, Any] = '' try: with open(UpperCAmelCase_ , 'rb' ) as binary_file: _UpperCamelCase : int = binary_file.read() for dat in data: _UpperCamelCase : List[str] = f'{dat:08b}' result += curr_byte return result except OSError: print('File not accessible' ) sys.exit() def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Any = {'0': '0', '1': '1'} _UpperCamelCase , _UpperCamelCase : int = '', '' _UpperCamelCase : List[str] = len(UpperCAmelCase_ ) for i in range(len(UpperCAmelCase_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue _UpperCamelCase : Union[str, Any] = lexicon[curr_string] result += last_match_id _UpperCamelCase : int = last_match_id + '0' if math.loga(UpperCAmelCase_ ).is_integer(): _UpperCamelCase : int = {} for curr_key in list(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = lexicon.pop(UpperCAmelCase_ ) _UpperCamelCase : Dict = new_lex _UpperCamelCase : Optional[int] = last_match_id + '1' index += 1 _UpperCamelCase : Optional[Any] = '' return result def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : str = 8 try: with open(UpperCAmelCase_ , 'wb' ) as opened_file: _UpperCamelCase : Dict = [ to_write[i : i + byte_length] for i in range(0 , len(UpperCAmelCase_ ) , UpperCAmelCase_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('10000000' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(UpperCAmelCase_ , 2 ).to_bytes(1 , byteorder='big' ) ) except OSError: print('File not accessible' ) sys.exit() def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = 0 for letter in data_bits: if letter == "1": break counter += 1 _UpperCamelCase : Dict = data_bits[counter:] _UpperCamelCase : Optional[Any] = data_bits[counter + 1 :] return data_bits def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[int] = read_file_binary(UpperCAmelCase_ ) _UpperCamelCase : Tuple = remove_prefix(UpperCAmelCase_ ) _UpperCamelCase : Dict = decompress_data(UpperCAmelCase_ ) write_file_binary(UpperCAmelCase_ , UpperCAmelCase_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
83
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase__ ( lowercase ): def __init__( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : str = dataset _UpperCamelCase : Optional[Any] = process _UpperCamelCase : Optional[Any] = params def __len__( self : Tuple ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.dataset[i] _UpperCamelCase : Dict = self.process(lowerCamelCase__ ,**self.params ) return processed class lowercase__ ( lowercase ): def __init__( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int]=None ): '''simple docstring''' _UpperCamelCase : Optional[int] = loader _UpperCamelCase : Tuple = infer _UpperCamelCase : List[str] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _UpperCamelCase : Any = None _UpperCamelCase : Union[str, Any] = loader_batch_size # Internal bookkeeping _UpperCamelCase : Optional[Any] = None _UpperCamelCase : str = None def __len__( self : List[str] ): '''simple docstring''' return len(self.loader ) def __iter__( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = iter(self.loader ) return self def UpperCamelCase_ ( self : Any ): '''simple docstring''' if isinstance(self._loader_batch_data ,torch.Tensor ): # Batch data is simple tensor, just fetch the slice _UpperCamelCase : Union[str, Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _UpperCamelCase : Union[str, Any] = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Convert ModelOutput to tuple first _UpperCamelCase : str = element.to_tuple() if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Dict = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _UpperCamelCase : Optional[int] = None elif isinstance(element[self._loader_batch_index] ,torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] ,np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : Optional[Any] = np.expand_dims(element[self._loader_batch_index] ,0 ) else: # This is typically a list, so no need to `unsqueeze`. _UpperCamelCase : Union[str, Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _UpperCamelCase : Optional[int] = self._loader_batch_data.__class__(lowerCamelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _UpperCamelCase : Tuple = next(self.iterator ) _UpperCamelCase : List[str] = self.infer(lowerCamelCase__ ,**self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : List[Any] = processed else: _UpperCamelCase : List[Any] = list(processed.keys() )[0] _UpperCamelCase : Optional[int] = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : int = len(lowerCamelCase__ ) else: _UpperCamelCase : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : int = observed_batch_size # Setting internal index to unwrap the batch _UpperCamelCase : Dict = processed _UpperCamelCase : str = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase__ ( lowercase ): def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __iter__( self : Dict ): '''simple docstring''' _UpperCamelCase : str = iter(self.loader ) _UpperCamelCase : List[str] = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: _UpperCamelCase : Tuple = self.infer(next(self.iterator ) ,**self.params ) try: # Try to return next item _UpperCamelCase : Optional[Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) _UpperCamelCase : int = next(self.subiterator ) return processed class lowercase__ ( lowercase ): def __iter__( self : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : Dict = self.loader_batch_item() _UpperCamelCase : List[str] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator while not is_last: _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) if self.loader_batch_size is not None: if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : str = processed else: _UpperCamelCase : Any = list(processed.keys() )[0] _UpperCamelCase : Tuple = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Dict = len(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : Any = observed_batch_size _UpperCamelCase : List[Any] = processed _UpperCamelCase : int = 0 while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : List[Any] = self.loader_batch_item() _UpperCamelCase : Optional[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator else: _UpperCamelCase : Any = processed _UpperCamelCase : List[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) return accumulator class lowercase__ ( lowercase ): def __init__( self : Tuple ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : str = key def __len__( self : Dict ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : Optional[Any] = keya _UpperCamelCase : str = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : List[str] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
83
1
'''simple docstring''' def A__ ( UpperCAmelCase_ = 2_0_0 ): _UpperCamelCase : str = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] _UpperCamelCase : List[Any] = [0] * (pence + 1) _UpperCamelCase : Any = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(UpperCAmelCase_ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
83
'''simple docstring''' import os from datetime import datetime as dt from github import Github snake_case_ : Any = [ 'good first issue', 'good second issue', 'good difficult issue', 'enhancement', 'new pipeline/model', 'new scheduler', 'wip', ] def A__ ( ): _UpperCamelCase : Tuple = Github(os.environ['GITHUB_TOKEN'] ) _UpperCamelCase : List[Any] = g.get_repo('huggingface/diffusers' ) _UpperCamelCase : List[Any] = repo.get_issues(state='open' ) for issue in open_issues: _UpperCamelCase : Dict = sorted(issue.get_comments() , key=lambda UpperCAmelCase_ : i.created_at , reverse=UpperCAmelCase_ ) _UpperCamelCase : List[str] = comments[0] if len(UpperCAmelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='closed' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='open' ) issue.remove_from_labels('stale' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) issue.add_to_labels('stale' ) if __name__ == "__main__": main()
83
1
'''simple docstring''' print((lambda quine: quine % quine)('print((lambda quine: quine %% quine)(%r))'))
83
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase ) , """Tatoeba directory does not exist.""" ) class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Dict = self.resolver.write_model_card('opus-mt-he-en' ,dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
1
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = len(UpperCAmelCase_ ) for i in range(1 , UpperCAmelCase_ ): _UpperCamelCase : str = collection[i] _UpperCamelCase : Optional[Any] = 0 _UpperCamelCase : List[Any] = i - 1 while low <= high: _UpperCamelCase : Any = (low + high) // 2 if val < collection[mid]: _UpperCamelCase : int = mid - 1 else: _UpperCamelCase : int = mid + 1 for j in range(UpperCAmelCase_ , UpperCAmelCase_ , -1 ): _UpperCamelCase : Any = collection[j - 1] _UpperCamelCase : List[Any] = val return collection if __name__ == "__main__": snake_case_ : Any = input('Enter numbers separated by a comma:\n').strip() snake_case_ : int = [int(item) for item in user_input.split(',')] print(binary_insertion_sort(unsorted))
83
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
1
'''simple docstring''' from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( """The RoBERTa Model transformer with early exiting (DeeRoBERTa). """ , lowercase , ) class lowercase__ ( lowercase ): lowercase__ = RobertaConfig lowercase__ = """roberta""" def __init__( self : str ,lowerCamelCase__ : Any ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : List[str] = RobertaEmbeddings(lowerCamelCase__ ) self.init_weights() @add_start_docstrings( """RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. """ , lowercase , ) class lowercase__ ( lowercase ): lowercase__ = RobertaConfig lowercase__ = """roberta""" def __init__( self : Optional[Any] ,lowerCamelCase__ : Any ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : Any = config.num_labels _UpperCamelCase : List[str] = config.num_hidden_layers _UpperCamelCase : int = DeeRobertaModel(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = nn.Dropout(config.hidden_dropout_prob ) _UpperCamelCase : Optional[Any] = nn.Linear(config.hidden_size ,self.config.num_labels ) @add_start_docstrings_to_model_forward(lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : List[Any]=None ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Tuple=None ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : Any=-1 ,lowerCamelCase__ : str=False ,): '''simple docstring''' _UpperCamelCase : Any = self.num_layers try: _UpperCamelCase : Tuple = self.roberta( lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,position_ids=lowerCamelCase__ ,head_mask=lowerCamelCase__ ,inputs_embeds=lowerCamelCase__ ,) _UpperCamelCase : Tuple = outputs[1] _UpperCamelCase : Optional[int] = self.dropout(lowerCamelCase__ ) _UpperCamelCase : List[str] = self.classifier(lowerCamelCase__ ) _UpperCamelCase : Tuple = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: _UpperCamelCase : Optional[int] = e.message _UpperCamelCase : List[str] = e.exit_layer _UpperCamelCase : Optional[Any] = outputs[0] if not self.training: _UpperCamelCase : int = entropy(lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = [] _UpperCamelCase : Optional[Any] = [] if labels is not None: if self.num_labels == 1: # We are doing regression _UpperCamelCase : Union[str, Any] = MSELoss() _UpperCamelCase : Tuple = loss_fct(logits.view(-1 ) ,labels.view(-1 ) ) else: _UpperCamelCase : int = CrossEntropyLoss() _UpperCamelCase : Dict = loss_fct(logits.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) # work with highway exits _UpperCamelCase : List[Any] = [] for highway_exit in outputs[-1]: _UpperCamelCase : List[Any] = highway_exit[0] if not self.training: highway_logits_all.append(lowerCamelCase__ ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression _UpperCamelCase : List[str] = MSELoss() _UpperCamelCase : Union[str, Any] = loss_fct(highway_logits.view(-1 ) ,labels.view(-1 ) ) else: _UpperCamelCase : Optional[Any] = CrossEntropyLoss() _UpperCamelCase : str = loss_fct(highway_logits.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) highway_losses.append(lowerCamelCase__ ) if train_highway: _UpperCamelCase : str = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: _UpperCamelCase : Union[str, Any] = (loss,) + outputs if not self.training: _UpperCamelCase : int = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: _UpperCamelCase : Any = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
83
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : Dict = 3 _UpperCamelCase : Any = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 1_5 == 0: result -= a a += 1 return result if __name__ == "__main__": print(F"""{solution() = }""")
83
1
'''simple docstring''' import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : Dict = ['a', 'b', 'c'] # Defaults to last layer if both are None _UpperCamelCase , _UpperCamelCase : Any = get_aligned_output_features_output_indices(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,['c'] ) self.assertEqual(lowerCamelCase__ ,[2] ) # Out indices set to match out features _UpperCamelCase , _UpperCamelCase : Union[str, Any] = get_aligned_output_features_output_indices(['a', 'c'] ,lowerCamelCase__ ,lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,['a', 'c'] ) self.assertEqual(lowerCamelCase__ ,[0, 2] ) # Out features set to match out indices _UpperCamelCase , _UpperCamelCase : List[Any] = get_aligned_output_features_output_indices(lowerCamelCase__ ,[0, 2] ,lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,['a', 'c'] ) self.assertEqual(lowerCamelCase__ ,[0, 2] ) # Out features selected from negative indices _UpperCamelCase , _UpperCamelCase : Any = get_aligned_output_features_output_indices(lowerCamelCase__ ,[-3, -1] ,lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ ,['a', 'c'] ) self.assertEqual(lowerCamelCase__ ,[-3, -1] ) def UpperCamelCase_ ( self : str ): '''simple docstring''' # Stage names must be set with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(['a', 'b'] ,(0, 1) ,lowerCamelCase__ ) # Out features must be a list with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(('a', 'b') ,(0, 1) ,['a', 'b'] ) # Out features must be a subset of stage names with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(['a', 'b'] ,(0, 1) ,['a'] ) # Out indices must be a list or tuple with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(lowerCamelCase__ ,0 ,['a', 'b'] ) # Out indices must be a subset of stage names with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(lowerCamelCase__ ,(0, 1) ,['a'] ) # Out features and out indices must be the same length with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(['a', 'b'] ,(0,) ,['a', 'b', 'c'] ) # Out features should match out indices with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(['a', 'b'] ,(0, 2) ,['a', 'b', 'c'] ) # Out features and out indices should be in order with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(['b', 'a'] ,(0, 1) ,['a', 'b'] ) # Check passes with valid inputs verify_out_features_out_indices(['a', 'b', 'd'] ,(0, 1, -1) ,['a', 'b', 'c', 'd'] ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : Dict = BackboneMixin() _UpperCamelCase : str = ['a', 'b', 'c'] _UpperCamelCase : Tuple = ['a', 'c'] _UpperCamelCase : Optional[int] = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,['a', 'c'] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly _UpperCamelCase : str = ['a', 'b'] self.assertEqual(backbone.out_features ,['a', 'b'] ) self.assertEqual(backbone.out_indices ,[0, 1] ) _UpperCamelCase : Union[str, Any] = [-3, -1] self.assertEqual(backbone.out_features ,['a', 'c'] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
83
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') snake_case_ : Any = logging.getLogger(__name__) @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) lowercase__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class lowercase__ : lowercase__ = field(default=lowercase , metadata={"""help""": """The input training data file (a text file)."""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def UpperCamelCase_ ( self : str ): '''simple docstring''' if self.train_file is not None: _UpperCamelCase : List[Any] = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _UpperCamelCase : Union[str, Any] = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowercase__ : lowercase__ = 42 lowercase__ = True lowercase__ = None lowercase__ = None def __call__( self : Optional[Any] ,lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = 'label' if 'label' in features[0].keys() else 'labels' _UpperCamelCase : List[Any] = [feature.pop(lowerCamelCase__ ) for feature in features] _UpperCamelCase : Dict = len(lowerCamelCase__ ) _UpperCamelCase : List[str] = len(features[0]['input_ids'] ) _UpperCamelCase : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(lowerCamelCase__ )] for feature in features ] _UpperCamelCase : str = list(chain(*lowerCamelCase__ ) ) _UpperCamelCase : Tuple = self.tokenizer.pad( lowerCamelCase__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='pt' ,) # Un-flatten _UpperCamelCase : str = {k: v.view(lowerCamelCase__ ,lowerCamelCase__ ,-1 ) for k, v in batch.items()} # Add back labels _UpperCamelCase : Optional[int] = torch.tensor(lowerCamelCase__ ,dtype=torch.intaa ) return batch def A__ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCamelCase : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Optional[int] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , UpperCAmelCase_ , UpperCAmelCase_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(UpperCAmelCase_ ) datasets.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCamelCase : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCamelCase : List[str] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _UpperCamelCase : Optional[int] = {} if data_args.train_file is not None: _UpperCamelCase : Tuple = data_args.train_file if data_args.validation_file is not None: _UpperCamelCase : Tuple = data_args.validation_file _UpperCamelCase : Any = data_args.train_file.split('.' )[-1] _UpperCamelCase : Union[str, Any] = load_dataset( UpperCAmelCase_ , data_files=UpperCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _UpperCamelCase : List[str] = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=UpperCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _UpperCamelCase : Any = [f'ending{i}' for i in range(4 )] _UpperCamelCase : int = 'sent1' _UpperCamelCase : List[str] = 'sent2' if data_args.max_seq_length is None: _UpperCamelCase : int = tokenizer.model_max_length if max_seq_length > 1_0_2_4: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) _UpperCamelCase : int = 1_0_2_4 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) _UpperCamelCase : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(UpperCAmelCase_ ): _UpperCamelCase : str = [[context] * 4 for context in examples[context_name]] _UpperCamelCase : Optional[Any] = examples[question_header_name] _UpperCamelCase : Tuple = [ [f'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(UpperCAmelCase_ ) ] # Flatten out _UpperCamelCase : Optional[int] = list(chain(*UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = list(chain(*UpperCAmelCase_ ) ) # Tokenize _UpperCamelCase : Tuple = tokenizer( UpperCAmelCase_ , UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(UpperCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _UpperCamelCase : Optional[Any] = raw_datasets['train'] if data_args.max_train_samples is not None: _UpperCamelCase : Tuple = min(len(UpperCAmelCase_ ) , data_args.max_train_samples ) _UpperCamelCase : Tuple = train_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): _UpperCamelCase : Union[str, Any] = train_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _UpperCamelCase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _UpperCamelCase : Union[str, Any] = min(len(UpperCAmelCase_ ) , data_args.max_eval_samples ) _UpperCamelCase : str = eval_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): _UpperCamelCase : Dict = eval_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _UpperCamelCase : List[Any] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=UpperCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Union[str, Any] = eval_predictions _UpperCamelCase : List[str] = np.argmax(UpperCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer _UpperCamelCase : Optional[int] = Trainer( model=UpperCAmelCase_ , args=UpperCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=UpperCAmelCase_ , data_collator=UpperCAmelCase_ , compute_metrics=UpperCAmelCase_ , ) # Training if training_args.do_train: _UpperCamelCase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _UpperCamelCase : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCamelCase : int = last_checkpoint _UpperCamelCase : List[str] = trainer.train(resume_from_checkpoint=UpperCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload _UpperCamelCase : Union[str, Any] = train_result.metrics _UpperCamelCase : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('train' , UpperCAmelCase_ ) trainer.save_metrics('train' , UpperCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _UpperCamelCase : List[Any] = trainer.evaluate() _UpperCamelCase : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(UpperCAmelCase_ ) _UpperCamelCase : int = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('eval' , UpperCAmelCase_ ) trainer.save_metrics('eval' , UpperCAmelCase_ ) _UpperCamelCase : Optional[int] = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**UpperCAmelCase_ ) else: trainer.create_model_card(**UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
83
1
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : str=3 ,lowerCamelCase__ : Union[str, Any]=32 ,lowerCamelCase__ : List[str]=3 ,lowerCamelCase__ : Union[str, Any]=10 ,lowerCamelCase__ : List[str]=[10, 20, 30, 40] ,lowerCamelCase__ : int=[1, 1, 2, 1] ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : List[str]="relu" ,lowerCamelCase__ : List[str]=3 ,lowerCamelCase__ : str=None ,): '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Tuple = batch_size _UpperCamelCase : Dict = image_size _UpperCamelCase : List[str] = num_channels _UpperCamelCase : Union[str, Any] = embeddings_size _UpperCamelCase : Union[str, Any] = hidden_sizes _UpperCamelCase : str = depths _UpperCamelCase : str = is_training _UpperCamelCase : Optional[int] = use_labels _UpperCamelCase : str = hidden_act _UpperCamelCase : Optional[int] = num_labels _UpperCamelCase : int = scope _UpperCamelCase : Union[str, Any] = len(lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCamelCase : List[str] = self.get_config() return config, pixel_values def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxRegNetModel(config=lowerCamelCase__ ) _UpperCamelCase : Optional[int] = model(lowerCamelCase__ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def UpperCamelCase_ ( self : str ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.num_labels _UpperCamelCase : Optional[int] = FlaxRegNetForImageClassification(config=lowerCamelCase__ ) _UpperCamelCase : Any = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : Tuple = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase : int = config_and_inputs _UpperCamelCase : List[str] = {'pixel_values': pixel_values} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase__ = False lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : str = FlaxRegNetModelTester(self ) _UpperCamelCase : Optional[int] = ConfigTester(self ,config_class=lowerCamelCase__ ,has_text_modality=lowerCamelCase__ ) def UpperCamelCase_ ( self : int ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' return def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @unittest.skip(reason='RegNet does not use inputs_embeds' ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='RegNet does not support input and output embeddings' ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' pass def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : Tuple = model_class(lowerCamelCase__ ) _UpperCamelCase : int = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase : List[Any] = [*signature.parameters.keys()] _UpperCamelCase : str = ['pixel_values'] self.assertListEqual(arg_names[:1] ,lowerCamelCase__ ) def UpperCamelCase_ ( self : int ): '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str] ): _UpperCamelCase : Union[str, Any] = model_class(lowerCamelCase__ ) _UpperCamelCase : int = model(**self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) ) _UpperCamelCase : int = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCamelCase : int = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) ,expected_num_stages + 1 ) _UpperCamelCase , _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : Any = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCamelCase : List[str] = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCamelCase : Any = self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : Dict = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ : List[Any] ,**lowerCamelCase__ : Any ): return model(pixel_values=lowerCamelCase__ ,**lowerCamelCase__ ) with self.subTest('JIT Enabled' ): _UpperCamelCase : Optional[int] = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): _UpperCamelCase : List[Any] = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) ,len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertEqual(jitted_output.shape ,output.shape ) def A__ ( ): _UpperCamelCase : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_flax class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : int ): '''simple docstring''' return AutoImageProcessor.from_pretrained('facebook/regnet-y-040' ) if is_vision_available() else None @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxRegNetForImageClassification.from_pretrained('facebook/regnet-y-040' ) _UpperCamelCase : Dict = self.default_image_processor _UpperCamelCase : str = prepare_img() _UpperCamelCase : Union[str, Any] = image_processor(images=lowerCamelCase__ ,return_tensors='np' ) _UpperCamelCase : Any = model(**lowerCamelCase__ ) # verify the logits _UpperCamelCase : List[Any] = (1, 1000) self.assertEqual(outputs.logits.shape ,lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = jnp.array([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] ,lowerCamelCase__ ,atol=1E-4 ) )
83
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) lowercase__ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) lowercase__ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def A__ ( ): _UpperCamelCase : Optional[Any] = HfArgumentParser((ModelArguments,) ) ((_UpperCamelCase) , ) : Optional[int] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _UpperCamelCase : Any = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _UpperCamelCase : List[Any] = True _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : str = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=UpperCAmelCase_ , decoder_config=UpperCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _UpperCamelCase : str = decoder_config.decoder_start_token_id _UpperCamelCase : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _UpperCamelCase : int = decoder_config.bos_token_id if pad_token_id is None: _UpperCamelCase : Dict = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _UpperCamelCase : List[Any] = decoder_config.eos_token_id _UpperCamelCase : Dict = decoder_start_token_id _UpperCamelCase : int = pad_token_id _UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
83
1
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if n == 1 or not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): return 0 elif n == 2: return 1 else: _UpperCamelCase : Tuple = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Dict = 0 _UpperCamelCase : str = 2 while digits < n: index += 1 _UpperCamelCase : Tuple = len(str(fibonacci(UpperCAmelCase_ ) ) ) return index def A__ ( UpperCAmelCase_ = 1_0_0_0 ): return fibonacci_digits_index(UpperCAmelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
83
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy snake_case_ : Dict = logging.get_logger(__name__) class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : float ,**lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : List[Any] = feature_size _UpperCamelCase : Any = sampling_rate _UpperCamelCase : Optional[Any] = padding_value _UpperCamelCase : Union[str, Any] = kwargs.pop('padding_side' ,'right' ) _UpperCamelCase : Dict = kwargs.pop('return_attention_mask' ,lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = True ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,): '''simple docstring''' # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowerCamelCase__ ,(list, tuple) ) and isinstance(processed_features[0] ,(dict, BatchFeature) ): _UpperCamelCase : int = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' F' to this method that includes {self.model_input_names[0]}, but you provided' F' {list(processed_features.keys() )}' ) _UpperCamelCase : List[Any] = processed_features[self.model_input_names[0]] _UpperCamelCase : Dict = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowerCamelCase__ ) == 0: if return_attention_mask: _UpperCamelCase : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch _UpperCamelCase : List[str] = required_input[0] if isinstance(lowerCamelCase__ ,(list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. _UpperCamelCase : List[str] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowerCamelCase__ ): _UpperCamelCase : Dict = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowerCamelCase__ ): _UpperCamelCase : Any = 'tf' elif is_torch_tensor(lowerCamelCase__ ): _UpperCamelCase : Optional[int] = 'pt' elif isinstance(lowerCamelCase__ ,(int, float, list, tuple, np.ndarray) ): _UpperCamelCase : int = 'np' else: raise ValueError( F'type of {first_element} unknown: {type(lowerCamelCase__ )}. ' 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] ,(int, float) ): _UpperCamelCase : Any = to_numpy(lowerCamelCase__ ) else: _UpperCamelCase : Any = [to_numpy(lowerCamelCase__ ) for v in value] # Convert padding_strategy in PaddingStrategy _UpperCamelCase : Optional[int] = self._get_padding_strategies(padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ) _UpperCamelCase : str = processed_features[self.model_input_names[0]] _UpperCamelCase : List[str] = len(lowerCamelCase__ ) if not all(len(lowerCamelCase__ ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) _UpperCamelCase : List[str] = [] for i in range(lowerCamelCase__ ): _UpperCamelCase : List[str] = {k: v[i] for k, v in processed_features.items()} # truncation _UpperCamelCase : List[str] = self._truncate( lowerCamelCase__ ,max_length=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,truncation=lowerCamelCase__ ,) truncated_inputs.append(lowerCamelCase__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length _UpperCamelCase : Union[str, Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) _UpperCamelCase : Any = PaddingStrategy.MAX_LENGTH _UpperCamelCase : Optional[Any] = {} for i in range(lowerCamelCase__ ): # padding _UpperCamelCase : Any = self._pad( truncated_inputs[i] ,max_length=lowerCamelCase__ ,padding_strategy=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,) for key, value in outputs.items(): if key not in batch_outputs: _UpperCamelCase : Dict = [] if value.dtype is np.dtype(np.floataa ): _UpperCamelCase : Any = value.astype(np.floataa ) batch_outputs[key].append(lowerCamelCase__ ) return BatchFeature(lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' _UpperCamelCase : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: _UpperCamelCase : Optional[Any] = len(lowerCamelCase__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : str = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : str = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowerCamelCase__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: _UpperCamelCase : Tuple = np.ones(len(lowerCamelCase__ ) ,dtype=np.intaa ) if needs_to_be_padded: _UpperCamelCase : Dict = max_length - len(lowerCamelCase__ ) if self.padding_side == "right": if return_attention_mask: _UpperCamelCase : Optional[int] = np.pad( processed_features['attention_mask'] ,(0, difference) ) _UpperCamelCase : Union[str, Any] = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) _UpperCamelCase : List[Any] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: _UpperCamelCase : List[Any] = np.pad( processed_features['attention_mask'] ,(difference, 0) ) _UpperCamelCase : List[Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) _UpperCamelCase : List[str] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) _UpperCamelCase : int = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : Optional[int] = len(lowerCamelCase__ ) > max_length if needs_to_be_truncated: _UpperCamelCase : Dict = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: _UpperCamelCase : Optional[Any] = processed_features['attention_mask'][:max_length] return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : int=False ,lowerCamelCase__ : Optional[Any]=None ): '''simple docstring''' # Get padding strategy if padding is not False: if padding is True: _UpperCamelCase : Optional[Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Tuple = PaddingStrategy(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = padding else: _UpperCamelCase : List[Any] = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
83
1
'''simple docstring''' import os import sys import unittest snake_case_ : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path snake_case_ : Tuple = os.path.join(git_repo_path, 'src', 'diffusers') class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : List[str] = find_backend(' if not is_torch_available():' ) self.assertEqual(lowerCamelCase__ ,'torch' ) # backend_with_underscore = find_backend(" if not is_tensorflow_text_available():") # self.assertEqual(backend_with_underscore, "tensorflow_text") _UpperCamelCase : Dict = find_backend(' if not (is_torch_available() and is_transformers_available()):' ) self.assertEqual(lowerCamelCase__ ,'torch_and_transformers' ) # double_backend_with_underscore = find_backend( # " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" # ) # self.assertEqual(double_backend_with_underscore, "sentencepiece_and_tensorflow_text") _UpperCamelCase : Optional[int] = find_backend( ' if not (is_torch_available() and is_transformers_available() and is_onnx_available()):' ) self.assertEqual(lowerCamelCase__ ,'torch_and_transformers_and_onnx' ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : str = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('torch' ,lowerCamelCase__ ) self.assertIn('torch_and_transformers' ,lowerCamelCase__ ) self.assertIn('flax_and_transformers' ,lowerCamelCase__ ) self.assertIn('torch_and_transformers_and_onnx' ,lowerCamelCase__ ) # Likewise, we can't assert on the exact content of a key self.assertIn('UNet2DModel' ,objects['torch'] ) self.assertIn('FlaxUNet2DConditionModel' ,objects['flax'] ) self.assertIn('StableDiffusionPipeline' ,objects['torch_and_transformers'] ) self.assertIn('FlaxStableDiffusionPipeline' ,objects['flax_and_transformers'] ) self.assertIn('LMSDiscreteScheduler' ,objects['torch_and_scipy'] ) self.assertIn('OnnxStableDiffusionPipeline' ,objects['torch_and_transformers_and_onnx'] ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : List[str] = create_dummy_object('CONSTANT' ,'\'torch\'' ) self.assertEqual(lowerCamelCase__ ,'\nCONSTANT = None\n' ) _UpperCamelCase : List[Any] = create_dummy_object('function' ,'\'torch\'' ) self.assertEqual( lowerCamelCase__ ,'\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n' ) _UpperCamelCase : str = '\nclass FakeClass(metaclass=DummyObject):\n _backends = \'torch\'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, \'torch\')\n\n @classmethod\n def from_config(cls, *args, **kwargs):\n requires_backends(cls, \'torch\')\n\n @classmethod\n def from_pretrained(cls, *args, **kwargs):\n requires_backends(cls, \'torch\')\n' _UpperCamelCase : Any = create_dummy_object('FakeClass' ,'\'torch\'' ) self.assertEqual(lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = '# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, ["torch"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = ["torch"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, ["torch"])\n\n @classmethod\n def from_config(cls, *args, **kwargs):\n requires_backends(cls, ["torch"])\n\n @classmethod\n def from_pretrained(cls, *args, **kwargs):\n requires_backends(cls, ["torch"])\n' _UpperCamelCase : int = create_dummy_files({'torch': ['CONSTANT', 'function', 'FakeClass']} ) self.assertEqual(dummy_files['torch'] ,lowerCamelCase__ )
83
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
1
from math import sqrt def _a ( a :int ) -> bool: assert isinstance(a , a ) and ( number >= 0 ), "'number' must been an int and positive" a = True # 0 and 1 are none primes. if number <= 1: a = False for divisor in range(2 , int(round(sqrt(a ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: a = False break # precondition assert isinstance(a , a ), "'status' must been from type bool" return status def _a ( a :Tuple ) -> str: assert isinstance(a , a ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N a = list(range(2 , n + 1 ) ) a = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(a ) ): for j in range(i + 1 , len(a ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): a = 0 # filters actual prime numbers. a = [x for x in begin_list if x != 0] # precondition assert isinstance(a , a ), "'ans' must been from type list" return ans def _a ( a :Optional[Any] ) -> Optional[Any]: assert isinstance(a , a ) and (n > 2), "'N' must been an int and > 2" a = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(a ): ans.append(a ) # precondition assert isinstance(a , a ), "'ans' must been from type list" return ans def _a ( a :Dict ) -> str: assert isinstance(a , a ) and number >= 0, "'number' must been an int and >= 0" a = [] # this list will be returns of the function. # potential prime number factors. a = 2 a = number if number == 0 or number == 1: ans.append(a ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(a ): while quotient != 1: if is_prime(a ) and (quotient % factor == 0): ans.append(a ) quotient /= factor else: factor += 1 else: ans.append(a ) # precondition assert isinstance(a , a ), "'ans' must been from type list" return ans def _a ( a :List[str] ) -> Union[str, Any]: assert isinstance(a , a ) and ( number >= 0 ), "'number' bust been an int and >= 0" a = 0 # prime factorization of 'number' a = prime_factorization(a ) a = max(a ) # precondition assert isinstance(a , a ), "'ans' must been from type int" return ans def _a ( a :Optional[int] ) -> List[str]: assert isinstance(a , a ) and ( number >= 0 ), "'number' bust been an int and >= 0" a = 0 # prime factorization of 'number' a = prime_factorization(a ) a = min(a ) # precondition assert isinstance(a , a ), "'ans' must been from type int" return ans def _a ( a :Any ) -> Any: assert isinstance(a , a ), "'number' must been an int" assert isinstance(number % 2 == 0 , a ), "compare bust been from type bool" return number % 2 == 0 def _a ( a :Any ) -> int: assert isinstance(a , a ), "'number' must been an int" assert isinstance(number % 2 != 0 , a ), "compare bust been from type bool" return number % 2 != 0 def _a ( a :Union[str, Any] ) -> Tuple: assert ( isinstance(a , a ) and (number > 2) and is_even(a ) ), "'number' must been an int, even and > 2" a = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' a = get_prime_numbers(a ) a = len(a ) # run variable for while-loops. a = 0 a = None # exit variable. for break up the loops a = True while i < len_pn and loop: a = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: a = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(a , a ) and (len(a ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def _a ( a :Dict , a :int ) -> Dict: assert ( isinstance(a , a ) and isinstance(a , a ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." a = 0 while numbera != 0: a = numbera % numbera a = numbera a = rest # precondition assert isinstance(a , a ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def _a ( a :str , a :str ) -> str: assert ( isinstance(a , a ) and isinstance(a , a ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." a = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' a = prime_factorization(a ) a = prime_factorization(a ) elif numbera == 1 or numbera == 1: a = [] a = [] a = max(a , a ) a = 0 a = 0 a = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: a = prime_fac_a.count(a ) a = prime_fac_a.count(a ) for _ in range(max(a , a ) ): ans *= n else: a = prime_fac_a.count(a ) for _ in range(a ): ans *= n done.append(a ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: a = prime_fac_a.count(a ) for _ in range(a ): ans *= n done.append(a ) # precondition assert isinstance(a , a ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def _a ( a :int ) -> Any: assert isinstance(a , a ) and (n >= 0), "'number' must been a positive int" a = 0 a = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(a ): ans += 1 # precondition assert isinstance(a , a ) and is_prime( a ), "'ans' must been a prime number and from type int" return ans def _a ( a :Optional[Any] , a :Optional[int] ) -> List[Any]: assert ( is_prime(a ) and is_prime(a ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" a = p_number_a + 1 # jump to the next number a = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(a ): number += 1 while number < p_number_a: ans.append(a ) number += 1 # fetch the next prime number. while not is_prime(a ): number += 1 # precondition assert ( isinstance(a , a ) and ans[0] != p_number_a and ans[len(a ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def _a ( a :Optional[Any] ) -> Optional[Any]: assert isinstance(a , a ) and (n >= 1), "'n' must been int and >= 1" a = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(a ) # precondition assert ans[0] == 1 and ans[len(a ) - 1] == n, "Error in function getDivisiors(...)" return ans def _a ( a :List[Any] ) -> Optional[Any]: assert isinstance(a , a ) and ( number > 1 ), "'number' must been an int and >= 1" a = get_divisors(a ) # precondition assert ( isinstance(a , a ) and (divisors[0] == 1) and (divisors[len(a ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def _a ( a :Optional[int] , a :Optional[int] ) -> Dict: assert ( isinstance(a , a ) and isinstance(a , a ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. a = gcd(abs(a ) , abs(a ) ) # precondition assert ( isinstance(a , a ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def _a ( a :str ) -> str: assert isinstance(a , a ) and (n >= 0), "'n' must been a int and >= 0" a = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def _a ( a :List[str] ) -> Optional[Any]: assert isinstance(a , a ) and (n >= 0), "'n' must been an int and >= 0" a = 0 a = 1 a = 1 # this will be return for _ in range(n - 1 ): a = ans ans += fiba a = tmp return ans
0
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowercase__ ( lowercase ): @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Dict = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : Dict = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : str = '1' _UpperCamelCase : Union[str, Any] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : List[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : List[Any] = self.get_env() _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # next emulate no network _UpperCamelCase : Dict = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : Dict = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : int = '\nfrom transformers import pipeline\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _UpperCamelCase : Union[str, Any] = self.get_env() _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Tuple = [sys.executable, '-c', '\n'.join([load, mock, run] )] _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,1 ,result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' ,result.stderr.decode().replace('\n' ,'' ) ,) @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = '\nfrom transformers import AutoModel\n ' _UpperCamelCase : int = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Any = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : Optional[int] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() )
83
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class __A ( metaclass=UpperCamelCase__ ): a__ : List[str] = ["""onnx"""] def __init__(self : List[Any] , *__a : Dict , **__a : Optional[Any] ): requires_backends(self , ["onnx"] ) @classmethod def _lowercase (cls : List[str] , *__a : Any , **__a : List[Any] ): requires_backends(cls , ["onnx"] ) @classmethod def _lowercase (cls : Optional[int] , *__a : Any , **__a : int ): requires_backends(cls , ["onnx"] )
1
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase__ ( unittest.TestCase ): def __init__( self : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=99 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=512 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : int=0.0_2 ,lowerCamelCase__ : int=4 ,): '''simple docstring''' _UpperCamelCase : List[Any] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Optional[Any] = is_training _UpperCamelCase : Optional[int] = use_attention_mask _UpperCamelCase : Any = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : str = intermediate_size _UpperCamelCase : int = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Optional[int] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_choices def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Union[str, Any] = None if self.use_attention_mask: _UpperCamelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=lowerCamelCase__ ,) return config, input_ids, attention_mask def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Optional[int] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase : Dict = model_class_name.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase : Tuple = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] _UpperCamelCase : Any = (1, 11, 768) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1E-4 ) )
83
0
'''simple docstring''' import os from collections.abc import Iterator def _SCREAMING_SNAKE_CASE (A = "." ) -> Iterator[str]: """simple docstring""" for dir_path, dir_names, filenames in os.walk(A ): lowercase__ = [d for d in dir_names if d != '''scripts''' and d[0] not in '''._'''] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A )[1] in (".py", ".ipynb"): yield os.path.join(A , A ).lstrip('''./''' ) def _SCREAMING_SNAKE_CASE (A ) -> List[str]: """simple docstring""" return f"{i * ' '}*" if i else "\n##" def _SCREAMING_SNAKE_CASE (A , A ) -> str: """simple docstring""" lowercase__ = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A ) or old_parts[i] != new_part) and new_part: print(f"{md_prefix(A )} {new_part.replace('_' , ' ' ).title()}" ) return new_path def _SCREAMING_SNAKE_CASE (A = "." ) -> None: """simple docstring""" lowercase__ = '''''' for filepath in sorted(good_file_paths(A ) ): lowercase__ ,lowercase__ = os.path.split(A ) if filepath != old_path: lowercase__ = print_path(A , A ) lowercase__ = (filepath.count(os.sep ) + 1) if filepath else 0 lowercase__ = f"{filepath}/{filename}".replace(''' ''' , '''%20''' ) lowercase__ = os.path.splitext(filename.replace('''_''' , ''' ''' ).title() )[0] print(f"{md_prefix(A )} [{filename}]({url})" ) if __name__ == "__main__": print_directory_md('.')
2
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer snake_case_ : List[Any] = logging.get_logger(__name__) class lowercase__ ( lowercase ): lowercase__ = """AutoTokenizer""" lowercase__ = ["""tokenizer"""] lowercase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self : List[str] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Tuple=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : Dict = speaker_embeddings @classmethod def UpperCamelCase_ ( cls : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : str="speaker_embeddings_path.json" ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: _UpperCamelCase : Optional[Any] = get_file_from_repo( lowerCamelCase__ ,lowerCamelCase__ ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(lowerCamelCase__ ,lowerCamelCase__ )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) _UpperCamelCase : Union[str, Any] = None else: with open(lowerCamelCase__ ) as speaker_embeddings_json: _UpperCamelCase : Optional[int] = json.load(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = None _UpperCamelCase : Tuple = AutoTokenizer.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) return cls(tokenizer=lowerCamelCase__ ,speaker_embeddings=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int="speaker_embeddings_path.json" ,lowerCamelCase__ : Dict="speaker_embeddings" ,lowerCamelCase__ : bool = False ,**lowerCamelCase__ : Tuple ,): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ,'v2' ) ,exist_ok=lowerCamelCase__ ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Optional[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCamelCase : Any = self._load_voice_preset(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] ,lowerCamelCase__ ,F'{prompt_key}_{key}' ) ,voice_preset[key] ,allow_pickle=lowerCamelCase__ ,) _UpperCamelCase : List[str] = os.path.join(lowerCamelCase__ ,F'{prompt_key}_{key}.npy' ) _UpperCamelCase : str = tmp_dict with open(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ) ,'w' ) as fp: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) super().save_pretrained(lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : str = None ,**lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.speaker_embeddings[voice_preset] _UpperCamelCase : Union[str, Any] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) _UpperCamelCase : Dict = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' ,'/' ) ,voice_preset_paths[key] ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" ,"/" ) ,voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) _UpperCamelCase : List[str] = np.load(lowerCamelCase__ ) return voice_preset_dict def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] ,np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self : Any ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any="pt" ,lowerCamelCase__ : Dict=256 ,lowerCamelCase__ : int=False ,lowerCamelCase__ : int=True ,lowerCamelCase__ : List[str]=False ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' if voice_preset is not None and not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if ( isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCamelCase : Optional[int] = self._load_voice_preset(lowerCamelCase__ ) else: if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and not voice_preset.endswith('.npz' ): _UpperCamelCase : Tuple = voice_preset + '.npz' _UpperCamelCase : str = np.load(lowerCamelCase__ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.tokenizer( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,padding='max_length' ,max_length=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,**lowerCamelCase__ ,) if voice_preset is not None: _UpperCamelCase : Optional[Any] = voice_preset return encoded_text
83
0
'''simple docstring''' import logging import os import threading import time try: import warnings except ImportError: lowercase : List[str] = None try: import msvcrt except ImportError: lowercase : Optional[int] = None try: import fcntl except ImportError: lowercase : Dict = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: lowercase : Tuple = OSError # Data # ------------------------------------------------ lowercase : Union[str, Any] = [ 'Timeout', 'BaseFileLock', 'WindowsFileLock', 'UnixFileLock', 'SoftFileLock', 'FileLock', ] lowercase : List[str] = '3.0.12' lowercase : Optional[Any] = None def lowerCAmelCase_ ( ): '''simple docstring''' global _logger A : Optional[Any] = _logger or logging.getLogger(__name__ ) return _logger class A ( __snake_case ): def __init__( self , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : str = lock_file return None def __str__( self ) -> List[Any]: """simple docstring""" A : Dict = F'The file lock \'{self.lock_file}\' could not be acquired.' return temp class A : def __init__( self , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" A : List[Any] = lock return None def __enter__( self ) -> Union[str, Any]: """simple docstring""" return self.lock def __exit__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" self.lock.release() return None class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=-1 , SCREAMING_SNAKE_CASE=None ) -> List[str]: """simple docstring""" A : str = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long A : Optional[Any] = self.hash_filename_if_too_long(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # The path to the lock file. A : Tuple = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. A : Tuple = None # The default timeout value. A : List[str] = timeout # We use this lock primarily for the lock counter. A : Optional[Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. A : Optional[Any] = 0 return None @property def __lowerCAmelCase ( self ) -> int: """simple docstring""" return self._lock_file @property def __lowerCAmelCase ( self ) -> int: """simple docstring""" return self._timeout @timeout.setter def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : Optional[Any] = float(SCREAMING_SNAKE_CASE ) return None def __lowerCAmelCase ( self ) -> int: """simple docstring""" raise NotImplementedError() def __lowerCAmelCase ( self ) -> Any: """simple docstring""" raise NotImplementedError() @property def __lowerCAmelCase ( self ) -> str: """simple docstring""" return self._lock_file_fd is not None def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=0.05 ) -> Optional[Any]: """simple docstring""" if timeout is None: A : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 A : List[str] = id(self ) A : Optional[Any] = self._lock_file A : Optional[Any] = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F'Attempting to acquire lock {lock_id} on {lock_filename}' ) self._acquire() if self.is_locked: logger().debug(F'Lock {lock_id} acquired on {lock_filename}' ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F'Timeout on acquiring lock {lock_id} on {lock_filename}' ) raise Timeout(self._lock_file ) else: logger().debug( F'Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...' ) time.sleep(SCREAMING_SNAKE_CASE ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: A : List[str] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE=False ) -> Optional[Any]: """simple docstring""" with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: A : List[Any] = id(self ) A : Tuple = self._lock_file logger().debug(F'Attempting to release lock {lock_id} on {lock_filename}' ) self._release() A : List[str] = 0 logger().debug(F'Lock {lock_id} released on {lock_filename}' ) return None def __enter__( self ) -> List[str]: """simple docstring""" self.acquire() return self def __exit__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" self.release() return None def __del__( self ) -> Any: """simple docstring""" self.release(force=SCREAMING_SNAKE_CASE ) return None def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : Optional[Any] = os.path.basename(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > max_length and max_length > 0: A : str = os.path.dirname(SCREAMING_SNAKE_CASE ) A : Optional[int] = str(hash(SCREAMING_SNAKE_CASE ) ) A : int = filename[: max_length - len(SCREAMING_SNAKE_CASE ) - 8] + '''...''' + hashed_filename + '''.lock''' return os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: return path class A ( __snake_case ): def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=-1 , SCREAMING_SNAKE_CASE=None ) -> Any: """simple docstring""" from .file_utils import relative_to_absolute_path super().__init__(SCREAMING_SNAKE_CASE , timeout=SCREAMING_SNAKE_CASE , max_filename_length=SCREAMING_SNAKE_CASE ) A : Optional[Any] = '''\\\\?\\''' + relative_to_absolute_path(self.lock_file ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : List[str] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: A : Tuple = os.open(self._lock_file , SCREAMING_SNAKE_CASE ) except OSError: pass else: try: msvcrt.locking(SCREAMING_SNAKE_CASE , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(SCREAMING_SNAKE_CASE ) else: A : Optional[Any] = fd return None def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : List[Any] = self._lock_file_fd A : Optional[Any] = None msvcrt.locking(SCREAMING_SNAKE_CASE , msvcrt.LK_UNLCK , 1 ) os.close(SCREAMING_SNAKE_CASE ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A ( __snake_case ): def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=-1 , SCREAMING_SNAKE_CASE=None ) -> Tuple: """simple docstring""" A : List[str] = os.statvfs(os.path.dirname(SCREAMING_SNAKE_CASE ) ).f_namemax super().__init__(SCREAMING_SNAKE_CASE , timeout=SCREAMING_SNAKE_CASE , max_filename_length=SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : int = os.O_RDWR | os.O_CREAT | os.O_TRUNC A : Tuple = os.open(self._lock_file , SCREAMING_SNAKE_CASE ) try: fcntl.flock(SCREAMING_SNAKE_CASE , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(SCREAMING_SNAKE_CASE ) else: A : Union[str, Any] = fd return None def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : Optional[int] = self._lock_file_fd A : str = None fcntl.flock(SCREAMING_SNAKE_CASE , fcntl.LOCK_UN ) os.close(SCREAMING_SNAKE_CASE ) return None class A ( __snake_case ): def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" A : Union[str, Any] = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: A : Tuple = os.open(self._lock_file , SCREAMING_SNAKE_CASE ) except OSError: pass else: A : List[Any] = fd return None def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" os.close(self._lock_file_fd ) A : Union[str, Any] = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None lowercase : Union[str, Any] = None if msvcrt: lowercase : Optional[int] = WindowsFileLock elif fcntl: lowercase : List[str] = UnixFileLock else: lowercase : Tuple = SoftFileLock if warnings is not None: warnings.warn('only soft file lock is available')
3
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin snake_case_ : Tuple = random.Random() def A__ ( UpperCAmelCase_ , UpperCAmelCase_=1.0 , UpperCAmelCase_=None , UpperCAmelCase_=None ): if rng is None: _UpperCamelCase : Dict = global_rng _UpperCamelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int=7 ,lowerCamelCase__ : str=400 ,lowerCamelCase__ : int=2000 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=16000 ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Optional[int]=True ,): '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Union[str, Any] = batch_size _UpperCamelCase : List[str] = min_seq_length _UpperCamelCase : Optional[int] = max_seq_length _UpperCamelCase : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _UpperCamelCase : List[str] = feature_size _UpperCamelCase : List[str] = padding_value _UpperCamelCase : List[Any] = sampling_rate _UpperCamelCase : Dict = return_attention_mask _UpperCamelCase : Tuple = do_normalize def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : Tuple=False ): '''simple docstring''' def _flatten(lowerCamelCase__ : Optional[Any] ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _UpperCamelCase : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _UpperCamelCase : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: _UpperCamelCase : int = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = WavaVecaFeatureExtractor def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): '''simple docstring''' self.assertTrue(np.all(np.mean(lowerCamelCase__ ,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ,axis=0 ) - 1 ) < 1E-3 ) ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' # Tests that all call wrap to encode_plus and batch_encode_plus _UpperCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _UpperCamelCase : int = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Tuple = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test not batched input _UpperCamelCase : Tuple = feat_extract(speech_inputs[0] ,return_tensors='np' ).input_values _UpperCamelCase : Any = feat_extract(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test batched _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : Optional[int] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _UpperCamelCase : str = [floats_list((1, x) )[0] for x in (800, 800, 800)] _UpperCamelCase : str = np.asarray(lowerCamelCase__ ) _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : int = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : str = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : List[str] = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,return_tensors='np' ) _UpperCamelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[str] = range(800 ,1400 ,200 ) _UpperCamelCase : List[str] = [floats_list((1, x) )[0] for x in lengths] _UpperCamelCase : Optional[Any] = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : str = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,max_length=lowerCamelCase__ ,padding=lowerCamelCase__ ) _UpperCamelCase : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Union[str, Any] = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='max_length' ,return_tensors='np' ) _UpperCamelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : int = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Any = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=2000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase_ ( self : Any ): '''simple docstring''' import torch _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = np.random.rand(100 ).astype(np.floataa ) _UpperCamelCase : Dict = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _UpperCamelCase : Optional[int] = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _UpperCamelCase : Tuple = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: _UpperCamelCase : Optional[int] = WavaVecaConfig.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase__ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask ,config.feat_extract_norm == 'layer' )
83
0
'''simple docstring''' import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class UpperCAmelCase_ ( ctypes.Structure ): # _fields is a specific attr expected by ctypes lowerCamelCase : List[Any] = [('''size''', ctypes.c_int), ('''visible''', ctypes.c_byte)] def a_ ( ): if os.name == "nt": lowerCAmelCase = CursorInfo() lowerCAmelCase = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) lowerCAmelCase = False ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def a_ ( ): if os.name == "nt": lowerCAmelCase = CursorInfo() lowerCAmelCase = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) lowerCAmelCase = True ctypes.windll.kernelaa.SetConsoleCursorInfo(lowerCamelCase , ctypes.byref(lowerCamelCase ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def a_ ( ): try: hide_cursor() yield finally: show_cursor()
4
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1 , UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : int = 1 _UpperCamelCase : Union[str, Any] = 0 for divide_by_number in range(UpperCAmelCase_ , digit + 1 ): _UpperCamelCase : list[int] = [] _UpperCamelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = len(UpperCAmelCase_ ) _UpperCamelCase : List[Any] = divide_by_number else: has_been_divided.append(UpperCAmelCase_ ) _UpperCamelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
83
0
UpperCAmelCase__ = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCAmelCase__ = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case ) -> list[int]: """simple docstring""" _lowercase =True _lowercase =[] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__snake_case , __snake_case , __snake_case ) order.append(__snake_case ) return order def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case ) -> list[int]: """simple docstring""" _lowercase =True _lowercase =[vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__snake_case , __snake_case , __snake_case ) return component def UpperCAmelCase_ ( __snake_case ) -> list[list[int]]: """simple docstring""" _lowercase =len(__snake_case ) * [False] _lowercase ={vert: [] for vert in range(len(__snake_case ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__snake_case ) _lowercase =[] for i, was_visited in enumerate(__snake_case ): if not was_visited: order += topology_sort(__snake_case , __snake_case , __snake_case ) _lowercase =[] _lowercase =len(__snake_case ) * [False] for i in range(len(__snake_case ) ): _lowercase =order[len(__snake_case ) - i - 1] if not visited[vert]: _lowercase =find_components(__snake_case , __snake_case , __snake_case ) components_list.append(__snake_case ) return components_list
5
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if num < 0: return False _UpperCamelCase : int = num _UpperCamelCase : int = 0 while num > 0: _UpperCamelCase : str = rev_num * 1_0 + (num % 1_0) num //= 1_0 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
83
0
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup A : str = logging.get_logger(__name__) class __A( a ): def __init__( self , **_snake_case ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''bs4'''] ) super().__init__(**_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> int: '''simple docstring''' __a = [] __a = [] __a = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag __a = parent.find_all(child.name , recursive=_snake_case ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(_snake_case ) else next(i for i, s in enumerate(_snake_case , 1 ) if s is child ) ) __a = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Optional[int]: '''simple docstring''' __a = BeautifulSoup(_snake_case , '''html.parser''' ) __a = [] __a = [] __a = [] for element in html_code.descendants: if type(_snake_case ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue __a = html.unescape(_snake_case ).strip() if not text_in_this_tag: continue all_doc_strings.append(_snake_case ) __a , __a = self.xpath_soup(_snake_case ) stringaxtag_seq.append(_snake_case ) stringaxsubs_seq.append(_snake_case ) if len(_snake_case ) != len(_snake_case ): raise ValueError('''Number of doc strings and xtags does not correspond''' ) if len(_snake_case ) != len(_snake_case ): raise ValueError('''Number of doc strings and xsubs does not correspond''' ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[int]: '''simple docstring''' __a = '''''' for tagname, subs in zip(_snake_case , _snake_case ): xpath += F"""/{tagname}""" if subs != 0: xpath += F"""[{subs}]""" return xpath def __call__( self , _snake_case ) -> BatchFeature: '''simple docstring''' __a = False # Check that strings has a valid type if isinstance(_snake_case , _snake_case ): __a = True elif isinstance(_snake_case , (list, tuple) ): if len(_snake_case ) == 0 or isinstance(html_strings[0] , _snake_case ): __a = True if not valid_strings: raise ValueError( '''HTML strings must of type `str`, `List[str]` (batch of examples), ''' F"""but is of type {type(_snake_case )}.""" ) __a = bool(isinstance(_snake_case , (list, tuple) ) and (isinstance(html_strings[0] , _snake_case )) ) if not is_batched: __a = [html_strings] # Get nodes + xpaths __a = [] __a = [] for html_string in html_strings: __a , __a , __a = self.get_three_from_single(_snake_case ) nodes.append(_snake_case ) __a = [] for node, tag_list, sub_list in zip(_snake_case , _snake_case , _snake_case ): __a = self.construct_xpath(_snake_case , _snake_case ) xpath_strings.append(_snake_case ) xpaths.append(_snake_case ) # return as Dict __a = {'''nodes''': nodes, '''xpaths''': xpaths} __a = BatchFeature(data=_snake_case , tensor_type=_snake_case ) return encoded_inputs
6
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = abs(UpperCAmelCase_ ) _UpperCamelCase : int = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = abs(UpperCAmelCase_ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def A__ ( UpperCAmelCase_ ): return sum(int(UpperCAmelCase_ ) for c in str(abs(UpperCAmelCase_ ) ) ) def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) -> None: _UpperCamelCase : str = f'{func.__name__}({value})' _UpperCamelCase : Tuple = timeit(f'__main__.{call}' , setup='import __main__' ) print(f'{call:56} = {func(UpperCAmelCase_ )} -- {timing:.4f} seconds' ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
83
0
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class A : """simple docstring""" lowerCamelCase = LEDConfig lowerCamelCase = {} lowerCamelCase = 'gelu' def __init__( self : str,lowercase_ : Any,lowercase_ : int=1_3,lowercase_ : Any=7,lowercase_ : List[str]=True,lowercase_ : Tuple=False,lowercase_ : List[Any]=9_9,lowercase_ : Dict=3_2,lowercase_ : Tuple=2,lowercase_ : str=4,lowercase_ : Optional[int]=3_7,lowercase_ : Optional[int]=0.1,lowercase_ : List[Any]=0.1,lowercase_ : List[str]=2_0,lowercase_ : List[str]=2,lowercase_ : Tuple=1,lowercase_ : List[Any]=0,lowercase_ : int=4,)-> Dict: '''simple docstring''' A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = eos_token_id A__ = pad_token_id A__ = bos_token_id A__ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after A__ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests A__ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def snake_case__ ( self : Dict )-> str: '''simple docstring''' A__ = ids_tensor([self.batch_size, self.seq_length - 1],self.vocab_size ) A__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ),1 ) A__ = tf.concat([input_ids, eos_tensor],axis=1 ) A__ = ids_tensor([self.batch_size, self.seq_length],self.vocab_size ) A__ = self.config_cls( vocab_size=self.vocab_size,d_model=self.hidden_size,encoder_layers=self.num_hidden_layers,decoder_layers=self.num_hidden_layers,encoder_attention_heads=self.num_attention_heads,decoder_attention_heads=self.num_attention_heads,encoder_ffn_dim=self.intermediate_size,decoder_ffn_dim=self.intermediate_size,dropout=self.hidden_dropout_prob,attention_dropout=self.attention_probs_dropout_prob,max_position_embeddings=self.max_position_embeddings,eos_token_ids=[2],bos_token_id=self.bos_token_id,pad_token_id=self.pad_token_id,decoder_start_token_id=self.pad_token_id,attention_window=self.attention_window,**self.config_updates,) A__ = prepare_led_inputs_dict(lowercase_,lowercase_,lowercase_ ) A__ = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]],axis=-1,) A__ = global_attention_mask return config, inputs_dict def snake_case__ ( self : Dict,lowercase_ : Optional[Any],lowercase_ : Any )-> int: '''simple docstring''' A__ = TFLEDModel(config=lowercase_ ).get_decoder() A__ = inputs_dict['input_ids'] A__ = input_ids[:1, :] A__ = inputs_dict['attention_mask'][:1, :] A__ = 1 # first forward pass A__ = model(lowercase_,attention_mask=lowercase_,use_cache=lowercase_ ) A__ , A__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids A__ = ids_tensor((self.batch_size, 3),config.vocab_size ) A__ = tf.cast(ids_tensor((self.batch_size, 3),2 ),tf.inta ) # append to next input_ids and A__ = tf.concat([input_ids, next_tokens],axis=-1 ) A__ = tf.concat([attention_mask, next_attn_mask],axis=-1 ) A__ = model(lowercase_,attention_mask=lowercase_ )[0] A__ = model(lowercase_,attention_mask=lowercase_,past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1],output_from_past.shape[1] ) # select random slice A__ = int(ids_tensor((1,),output_from_past.shape[-1] ) ) A__ = output_from_no_past[:, -3:, random_slice_idx] A__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_,lowercase_,rtol=1E-3 ) def _snake_case( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , SCREAMING_SNAKE_CASE__ : Any=None , SCREAMING_SNAKE_CASE__ : List[Any]=None , SCREAMING_SNAKE_CASE__ : Optional[int]=None , ) -> Any: '''simple docstring''' if attention_mask is None: A__ = tf.cast(tf.math.not_equal(SCREAMING_SNAKE_CASE__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: A__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: A__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: A__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class A ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) lowerCamelCase = True lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def snake_case__ ( self : Any )-> Optional[Any]: '''simple docstring''' A__ = TFLEDModelTester(self ) A__ = ConfigTester(self,config_class=lowercase_ ) def snake_case__ ( self : Dict )-> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def snake_case__ ( self : Optional[Any] )-> Union[str, Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def snake_case__ ( self : List[Any] )-> Union[str, Any]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = tf.zeros_like(inputs_dict['attention_mask'] ) A__ = 2 A__ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices,1,inputs_dict['global_attention_mask'],) A__ = True A__ = self.model_tester.seq_length A__ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ : int ): A__ = outputs.decoder_attentions self.assertEqual(len(lowercase_ ),self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ),[self.model_tester.num_attention_heads, seq_length, seq_length],) def check_encoder_attentions_output(lowercase_ : List[str] ): A__ = [t.numpy() for t in outputs.encoder_attentions] A__ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ),self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ),self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ),[self.model_tester.num_attention_heads, seq_length, seq_length],) self.assertListEqual( list(global_attentions[0].shape[-3:] ),[self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices],) for model_class in self.all_model_classes: A__ = True A__ = False A__ = False A__ = model_class(lowercase_ ) A__ = model(self._prepare_for_class(lowercase_,lowercase_ ) ) A__ = len(lowercase_ ) self.assertEqual(config.output_hidden_states,lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: A__ = model_class(lowercase_ ) A__ = model(self._prepare_for_class(lowercase_,lowercase_ ) ) self.assertEqual(config.output_hidden_states,lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] A__ = True A__ = model_class(lowercase_ ) A__ = model(self._prepare_for_class(lowercase_,lowercase_ ) ) self.assertEqual(config.output_hidden_states,lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine A__ = True A__ = True A__ = model_class(lowercase_ ) A__ = model(self._prepare_for_class(lowercase_,lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1),len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states,lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def snake_case__ ( self : Optional[int] )-> Dict: '''simple docstring''' pass def snake_case__ ( self : Union[str, Any] )-> Optional[Any]: '''simple docstring''' pass def _snake_case( SCREAMING_SNAKE_CASE__ : Dict ) -> Any: '''simple docstring''' return tf.constant(SCREAMING_SNAKE_CASE__ , dtype=tf.intaa ) lowercase_ = 1e-4 @slow @require_tf class A ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : str )-> Union[str, Any]: '''simple docstring''' A__ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here A__ = _long_tensor([5_1_2 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) A__ = _long_tensor([1_2_8 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) A__ = prepare_led_inputs_dict(model.config,lowercase_,lowercase_ ) A__ = model(**lowercase_ )[0] A__ = (1, 1_0_2_4, 7_6_8) self.assertEqual(output.shape,lowercase_ ) # change to expected output here A__ = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]],) tf.debugging.assert_near(output[:, :3, :3],lowercase_,atol=1E-3 ) def snake_case__ ( self : Union[str, Any] )-> Optional[Any]: '''simple docstring''' A__ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here A__ = _long_tensor([5_1_2 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) A__ = _long_tensor([1_2_8 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) A__ = prepare_led_inputs_dict(model.config,lowercase_,lowercase_ ) A__ = model(**lowercase_ )[0] A__ = (1, 1_0_2_4, model.config.vocab_size) self.assertEqual(output.shape,lowercase_ ) # change to expected output here A__ = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]],) tf.debugging.assert_near(output[:, :3, :3],lowercase_,atol=1E-3,rtol=1E-3 )
7
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
0
import numpy as np from transformers import Pipeline def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): snake_case_ = np.max(SCREAMING_SNAKE_CASE__ , axis=-1 , keepdims=SCREAMING_SNAKE_CASE__ ) snake_case_ = np.exp(outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=SCREAMING_SNAKE_CASE__ ) class snake_case_ ( __A ): '''simple docstring''' def snake_case__( self : Tuple , **_UpperCamelCase : Tuple ) ->Optional[int]: snake_case_ = {} if "second_text" in kwargs: snake_case_ = kwargs['''second_text'''] return preprocess_kwargs, {}, {} def snake_case__( self : int , _UpperCamelCase : int , _UpperCamelCase : List[Any]=None ) ->Tuple: return self.tokenizer(_UpperCamelCase , text_pair=_UpperCamelCase , return_tensors=self.framework ) def snake_case__( self : List[str] , _UpperCamelCase : int ) ->str: return self.model(**_UpperCamelCase ) def snake_case__( self : str , _UpperCamelCase : List[Any] ) ->Tuple: snake_case_ = model_outputs.logits[0].numpy() snake_case_ = softmax(_UpperCamelCase ) snake_case_ = np.argmax(_UpperCamelCase ) snake_case_ = self.model.config.idalabel[best_class] snake_case_ = probabilities[best_class].item() snake_case_ = logits.tolist() return {"label": label, "score": score, "logits": logits}
8
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
0
from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def _UpperCamelCase ( lowercase__ ): return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) __lowerCAmelCase : Optional[Any] ='\ntransformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires\nTensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions.\n' class _lowercase ( A__ ): '''simple docstring''' @staticmethod def __magic_name__( lowerCAmelCase__ :ArgumentParser ) -> Any: __SCREAMING_SNAKE_CASE : List[Any] = parser.add_parser( '''convert''' , help='''CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints.''' , ) train_parser.add_argument('''--model_type''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Model\'s type.''' ) train_parser.add_argument( '''--tf_checkpoint''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''TensorFlow checkpoint path or folder.''' ) train_parser.add_argument( '''--pytorch_dump_output''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Path to the PyTorch saved model output.''' ) train_parser.add_argument('''--config''' , type=lowerCAmelCase__ , default='''''' , help='''Configuration file path or folder.''' ) train_parser.add_argument( '''--finetuning_task_name''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Optional fine-tuning task name if the TF model was a finetuned model.''' , ) train_parser.set_defaults(func=lowerCAmelCase__ ) def __init__( self :int , lowerCAmelCase__ :str , lowerCAmelCase__ :str , lowerCAmelCase__ :str , lowerCAmelCase__ :str , lowerCAmelCase__ :str , *lowerCAmelCase__ :List[Any] , ) -> Dict: __SCREAMING_SNAKE_CASE : str = logging.get_logger('''transformers-cli/converting''' ) self._logger.info(f'''Loading model {model_type}''' ) __SCREAMING_SNAKE_CASE : Optional[int] = model_type __SCREAMING_SNAKE_CASE : str = tf_checkpoint __SCREAMING_SNAKE_CASE : Optional[int] = pytorch_dump_output __SCREAMING_SNAKE_CASE : str = config __SCREAMING_SNAKE_CASE : Optional[int] = finetuning_task_name def __magic_name__( self :int ) -> List[str]: if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) if "ckpt" in self._tf_checkpoint.lower(): __SCREAMING_SNAKE_CASE : Dict = self._tf_checkpoint __SCREAMING_SNAKE_CASE : int = '''''' else: __SCREAMING_SNAKE_CASE : int = self._tf_checkpoint __SCREAMING_SNAKE_CASE : List[Any] = '''''' convert_transfo_xl_checkpoint_to_pytorch( lowerCAmelCase__ , self._config , self._pytorch_dump_output , lowerCAmelCase__ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( '''--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]''' )
9
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
0
from random import randint, random def lowerCAmelCase_ ( __a , __a , __a , __a = False , __a = False , __a = 5 , ) -> list: """simple docstring""" lowerCamelCase__: Tuple =[[-1] * number_of_cells] # Create a highway without any car lowerCamelCase__: Dict =0 lowerCamelCase__: Any =max(__a , 0 ) while i < number_of_cells: lowerCamelCase__: str =( randint(0 , __a ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def lowerCAmelCase_ ( __a , __a ) -> int: """simple docstring""" lowerCamelCase__: Optional[int] =0 lowerCamelCase__: Optional[int] =highway_now[car_index + 1 :] for cell in range(len(__a ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(__a , -1 ) def lowerCAmelCase_ ( __a , __a , __a ) -> list: """simple docstring""" lowerCamelCase__: Optional[int] =len(__a ) # Beforce calculations, the highway is empty lowerCamelCase__: Dict =[-1] * number_of_cells for car_index in range(__a ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed lowerCamelCase__: int =min(highway_now[car_index] + 1 , __a ) # Number of empty cell before the next car lowerCamelCase__: Union[str, Any] =get_distance(__a , __a ) - 1 # We can't have the car causing an accident lowerCamelCase__: Dict =min(next_highway[car_index] , __a ) if random() < probability: # Randomly, a driver will slow down lowerCamelCase__: int =max(next_highway[car_index] - 1 , 0 ) return next_highway def lowerCAmelCase_ ( __a , __a , __a , __a ) -> list: """simple docstring""" lowerCamelCase__: Dict =len(highway[0] ) for i in range(__a ): lowerCamelCase__: Dict =update(highway[i] , __a , __a ) lowerCamelCase__: List[str] =[-1] * number_of_cells for car_index in range(__a ): lowerCamelCase__: Any =next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) lowerCamelCase__: Dict =(car_index + speed) % number_of_cells # Commit the change of position lowerCamelCase__: Optional[Any] =speed highway.append(__a ) return highway if __name__ == "__main__": import doctest doctest.testmod()
10
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
0
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase__ ( a): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase = False , __lowerCamelCase = False , __lowerCamelCase = None , __lowerCamelCase = None , **__lowerCamelCase , ) -> Optional[int]: super().__init__( __lowerCamelCase , split=__lowerCamelCase , features=__lowerCamelCase , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase , streaming=__lowerCamelCase , num_proc=__lowerCamelCase , **__lowerCamelCase , ) _A : str = field _A : Dict = path_or_paths if isinstance(__lowerCamelCase , __lowerCamelCase) else {self.split: path_or_paths} _A : Tuple = Json( cache_dir=__lowerCamelCase , data_files=__lowerCamelCase , features=__lowerCamelCase , field=__lowerCamelCase , **__lowerCamelCase , ) def _lowerCamelCase ( self) -> Optional[Any]: # Build iterable dataset if self.streaming: _A : Optional[Any] = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _A : List[Any] = None _A : Any = None _A : Union[str, Any] = None _A : Dict = None self.builder.download_and_prepare( download_config=__lowerCamelCase , download_mode=__lowerCamelCase , verification_mode=__lowerCamelCase , base_path=__lowerCamelCase , num_proc=self.num_proc , ) _A : Any = self.builder.as_dataset( split=self.split , verification_mode=__lowerCamelCase , in_memory=self.keep_in_memory) return dataset class lowerCAmelCase__ : '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = None , __lowerCamelCase = None , **__lowerCamelCase , ) -> Optional[int]: if num_proc is not None and num_proc <= 0: raise ValueError(F"num_proc {num_proc} must be an integer > 0.") _A : Optional[int] = dataset _A : Optional[Any] = path_or_buf _A : Any = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE _A : Dict = num_proc _A : Optional[Any] = "utf-8" _A : str = to_json_kwargs def _lowerCamelCase ( self) -> int: _A : Tuple = self.to_json_kwargs.pop("path_or_buf" , __lowerCamelCase) _A : int = self.to_json_kwargs.pop("orient" , "records") _A : Any = self.to_json_kwargs.pop("lines" , True if orient == "records" else False) _A : Optional[Any] = self.to_json_kwargs.pop("index" , False if orient in ["split", "table"] else True) _A : List[Any] = self.to_json_kwargs.pop("compression" , __lowerCamelCase) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F"`datasets` currently does not support {compression} compression") if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with fsspec.open(self.path_or_buf , "wb" , compression=__lowerCamelCase) as buffer: _A : int = self._write(file_obj=__lowerCamelCase , orient=__lowerCamelCase , lines=__lowerCamelCase , index=__lowerCamelCase , **self.to_json_kwargs) else: if compression: raise NotImplementedError( F"The compression parameter is not supported when writing to a buffer, but compression={compression}" " was passed. Please provide a local path instead.") _A : Dict = self._write( file_obj=self.path_or_buf , orient=__lowerCamelCase , lines=__lowerCamelCase , index=__lowerCamelCase , **self.to_json_kwargs) return written def _lowerCamelCase ( self , __lowerCamelCase) -> Union[str, Any]: _A , _A , _A , _A , _A : str = args _A : Any = query_table( table=self.dataset.data , key=slice(__lowerCamelCase , offset + self.batch_size) , indices=self.dataset._indices , ) _A : Optional[int] = batch.to_pandas().to_json( path_or_buf=__lowerCamelCase , orient=__lowerCamelCase , lines=__lowerCamelCase , index=__lowerCamelCase , **__lowerCamelCase) if not json_str.endswith("\n"): json_str += "\n" return json_str.encode(self.encoding) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , **__lowerCamelCase , ) -> int: _A : Any = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset) , self.batch_size) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): _A : Any = self._batch_json((offset, orient, lines, index, to_json_kwargs)) written += file_obj.write(__lowerCamelCase) else: _A , _A : List[str] = len(self.dataset), self.batch_size with multiprocessing.Pool(self.num_proc) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , __lowerCamelCase , __lowerCamelCase)] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): written += file_obj.write(__lowerCamelCase) return written
11
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
0
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase__( __lowerCamelCase): def __init__( self: str , UpperCamelCase_: int , UpperCamelCase_: List[Any] , UpperCamelCase_: str=10_24 , UpperCamelCase_: int=10_24 , UpperCamelCase_: Any=3.6 ): __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self: Optional[int] ): __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase, __lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(UpperCamelCase_ )["""content"""] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(UpperCamelCase_ , truncation=UpperCamelCase_ )["""input_ids"""] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(UpperCamelCase_ ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(UpperCamelCase_ ) == self.seq_length: yield torch.tensor(UpperCamelCase_ ) def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = {"""streaming""": True} __lowerCamelCase = load_dataset(args.dataset_name , split="""train""" , **A__ ) __lowerCamelCase = ConstantLengthDataset(A__ , A__ , seq_length=args.seq_length ) __lowerCamelCase = DataLoader(A__ , batch_size=args.batch_size ) return eval_dataloader def lowerCamelCase__ ( A__ : List[str] ): '''simple docstring''' model.eval() __lowerCamelCase = [] for step, batch in enumerate(A__ ): with torch.no_grad(): __lowerCamelCase = model(A__ , labels=A__ ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(A__ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(A__ ) ) try: __lowerCamelCase = torch.exp(A__ ) except OverflowError: __lowerCamelCase = float("""inf""" ) return loss.item(), perplexity.item() # Setup Accelerator UpperCAmelCase_ = Accelerator() # Parse configuration UpperCAmelCase_ = HfArgumentParser(EvaluationArguments) UpperCAmelCase_ = parser.parse_args() set_seed(args.seed) # Logging UpperCAmelCase_ = logging.getLogger(__name__) logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) # Load model and tokenizer UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) UpperCAmelCase_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader UpperCAmelCase_ = create_dataloader(args) # Prepare everything with our `accelerator`. UpperCAmelCase_ , UpperCAmelCase_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('Evaluating and saving model after training') UpperCAmelCase_ , UpperCAmelCase_ = evaluate(args) logger.info(f"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
12
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase__ ( lowercase ): def __init__( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : str = dataset _UpperCamelCase : Optional[Any] = process _UpperCamelCase : Optional[Any] = params def __len__( self : Tuple ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.dataset[i] _UpperCamelCase : Dict = self.process(lowerCamelCase__ ,**self.params ) return processed class lowercase__ ( lowercase ): def __init__( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int]=None ): '''simple docstring''' _UpperCamelCase : Optional[int] = loader _UpperCamelCase : Tuple = infer _UpperCamelCase : List[str] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _UpperCamelCase : Any = None _UpperCamelCase : Union[str, Any] = loader_batch_size # Internal bookkeeping _UpperCamelCase : Optional[Any] = None _UpperCamelCase : str = None def __len__( self : List[str] ): '''simple docstring''' return len(self.loader ) def __iter__( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = iter(self.loader ) return self def UpperCamelCase_ ( self : Any ): '''simple docstring''' if isinstance(self._loader_batch_data ,torch.Tensor ): # Batch data is simple tensor, just fetch the slice _UpperCamelCase : Union[str, Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _UpperCamelCase : Union[str, Any] = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Convert ModelOutput to tuple first _UpperCamelCase : str = element.to_tuple() if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Dict = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _UpperCamelCase : Optional[int] = None elif isinstance(element[self._loader_batch_index] ,torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] ,np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : Optional[Any] = np.expand_dims(element[self._loader_batch_index] ,0 ) else: # This is typically a list, so no need to `unsqueeze`. _UpperCamelCase : Union[str, Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _UpperCamelCase : Optional[int] = self._loader_batch_data.__class__(lowerCamelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _UpperCamelCase : Tuple = next(self.iterator ) _UpperCamelCase : List[str] = self.infer(lowerCamelCase__ ,**self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : List[Any] = processed else: _UpperCamelCase : List[Any] = list(processed.keys() )[0] _UpperCamelCase : Optional[int] = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : int = len(lowerCamelCase__ ) else: _UpperCamelCase : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : int = observed_batch_size # Setting internal index to unwrap the batch _UpperCamelCase : Dict = processed _UpperCamelCase : str = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase__ ( lowercase ): def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __iter__( self : Dict ): '''simple docstring''' _UpperCamelCase : str = iter(self.loader ) _UpperCamelCase : List[str] = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: _UpperCamelCase : Tuple = self.infer(next(self.iterator ) ,**self.params ) try: # Try to return next item _UpperCamelCase : Optional[Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) _UpperCamelCase : int = next(self.subiterator ) return processed class lowercase__ ( lowercase ): def __iter__( self : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : Dict = self.loader_batch_item() _UpperCamelCase : List[str] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator while not is_last: _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) if self.loader_batch_size is not None: if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : str = processed else: _UpperCamelCase : Any = list(processed.keys() )[0] _UpperCamelCase : Tuple = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Dict = len(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : Any = observed_batch_size _UpperCamelCase : List[Any] = processed _UpperCamelCase : int = 0 while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : List[Any] = self.loader_batch_item() _UpperCamelCase : Optional[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator else: _UpperCamelCase : Any = processed _UpperCamelCase : List[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) return accumulator class lowercase__ ( lowercase ): def __init__( self : Tuple ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : str = key def __len__( self : Dict ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : Optional[Any] = keya _UpperCamelCase : str = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : List[str] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
83
0
import pprint import requests lowerCAmelCase : Dict = """https://zenquotes.io/api""" def A_ ( ): return requests.get(API_ENDPOINT_URL + "/today" ).json() def A_ ( ): return requests.get(API_ENDPOINT_URL + "/random" ).json() if __name__ == "__main__": lowerCAmelCase : List[str] = random_quotes() pprint.pprint(response)
13
'''simple docstring''' import os from datetime import datetime as dt from github import Github snake_case_ : Any = [ 'good first issue', 'good second issue', 'good difficult issue', 'enhancement', 'new pipeline/model', 'new scheduler', 'wip', ] def A__ ( ): _UpperCamelCase : Tuple = Github(os.environ['GITHUB_TOKEN'] ) _UpperCamelCase : List[Any] = g.get_repo('huggingface/diffusers' ) _UpperCamelCase : List[Any] = repo.get_issues(state='open' ) for issue in open_issues: _UpperCamelCase : Dict = sorted(issue.get_comments() , key=lambda UpperCAmelCase_ : i.created_at , reverse=UpperCAmelCase_ ) _UpperCamelCase : List[str] = comments[0] if len(UpperCAmelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='closed' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='open' ) issue.remove_from_labels('stale' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) issue.add_to_labels('stale' ) if __name__ == "__main__": main()
83
0
import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) _lowerCamelCase : List[Any] = """bert-base-cased""" _lowerCamelCase : Any = """fp16""" _lowerCamelCase : Tuple = """bf16""" _lowerCamelCase : List[str] = [FPaa, BFaa] @require_fsdp @require_cuda class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : List[Any]) ->List[Any]: '''simple docstring''' super().setUp() A__ = dict( ACCELERATE_USE_FSDP='''true''' , MASTER_ADDR='''localhost''' , MASTER_PORT='''10999''' , RANK='''0''' , LOCAL_RANK='''0''' , WORLD_SIZE='''1''' , ) def SCREAMING_SNAKE_CASE ( self : str) ->List[str]: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(UpperCAmelCase__): A__ = self.dist_env.copy() A__ = f"""{i + 1}""" A__ = strategy with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1)) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Tuple: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(UpperCAmelCase__): A__ = self.dist_env.copy() A__ = prefetch_policy with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1)) def SCREAMING_SNAKE_CASE ( self : Tuple) ->Dict: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(UpperCAmelCase__): A__ = self.dist_env.copy() A__ = state_dict_type with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1)) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only) def SCREAMING_SNAKE_CASE ( self : str) ->int: '''simple docstring''' A__ = AutoModel.from_pretrained(UpperCAmelCase__) for policy in FSDP_AUTO_WRAP_POLICY: A__ = self.dist_env.copy() A__ = policy if policy == "TRANSFORMER_BASED_WRAP": A__ = '''BertLayer''' elif policy == "SIZE_BASED_WRAP": A__ = '''2000''' with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(UpperCAmelCase__) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy) A__ = self.dist_env.copy() A__ = '''TRANSFORMER_BASED_WRAP''' A__ = '''T5Layer''' with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() with self.assertRaises(UpperCAmelCase__) as cm: fsdp_plugin.set_auto_wrap_policy(UpperCAmelCase__) self.assertTrue('''Could not find the transformer layer class to wrap in the model.''' in str(cm.exception)) A__ = self.dist_env.copy() A__ = '''SIZE_BASED_WRAP''' A__ = '''0''' with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(UpperCAmelCase__) self.assertIsNone(fsdp_plugin.auto_wrap_policy) def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Optional[int]: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: A__ = self.dist_env.copy() A__ = mp_dtype with mockenv_context(**UpperCAmelCase__): A__ = Accelerator() if mp_dtype == "fp16": A__ = torch.floataa elif mp_dtype == "bf16": A__ = torch.bfloataa A__ = MixedPrecision(param_dtype=UpperCAmelCase__ , reduce_dtype=UpperCAmelCase__ , buffer_dtype=UpperCAmelCase__) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , UpperCAmelCase__) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , UpperCAmelCase__)) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler) AcceleratorState._reset_state(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Dict) ->str: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: A__ = self.dist_env.copy() A__ = str(UpperCAmelCase__).lower() with mockenv_context(**UpperCAmelCase__): A__ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=UpperCAmelCase__)) @require_fsdp @require_multi_gpu @slow class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : str) ->List[Any]: '''simple docstring''' super().setUp() A__ = 0.82 A__ = [ '''fsdp_shard_grad_op_transformer_based_wrap''', '''fsdp_full_shard_transformer_based_wrap''', ] A__ = { '''multi_gpu_fp16''': 3_200, '''fsdp_shard_grad_op_transformer_based_wrap_fp16''': 2_000, '''fsdp_full_shard_transformer_based_wrap_fp16''': 1_900, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } A__ = 160 A__ = 160 A__ = inspect.getfile(accelerate.test_utils) A__ = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''external_deps''']) def SCREAMING_SNAKE_CASE ( self : str) ->Any: '''simple docstring''' A__ = os.path.join(self.test_scripts_folder , '''test_performance.py''') A__ = ['''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', '''--use_fsdp'''] for config in self.performance_configs: A__ = cmd.copy() for i, strategy in enumerate(UpperCAmelCase__): if strategy.lower() in config: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""") break if "fp32" in config: cmd_config.append('''--mixed_precision=no''') else: cmd_config.append('''--mixed_precision=fp16''') if "cpu_offload" in config: cmd_config.append('''--fsdp_offload_params=True''') for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""") break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''') elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''') cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--performance_lower_bound={self.performance_lower_bound}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(UpperCAmelCase__ , env=os.environ.copy()) def SCREAMING_SNAKE_CASE ( self : List[str]) ->Optional[Any]: '''simple docstring''' A__ = os.path.join(self.test_scripts_folder , '''test_checkpointing.py''') A__ = [ '''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', '''--use_fsdp''', '''--mixed_precision=fp16''', '''--fsdp_transformer_layer_cls_to_wrap=BertLayer''', ] for i, strategy in enumerate(UpperCAmelCase__): A__ = cmd.copy() cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""") if strategy != "FULL_SHARD": continue A__ = len(UpperCAmelCase__) for state_dict_type in FSDP_STATE_DICT_TYPE: A__ = cmd_config[:state_dict_config_index] cmd_config.append(f"""--fsdp_state_dict_type={state_dict_type}""") cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", '''--partial_train_epoch=1''', ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(UpperCAmelCase__ , env=os.environ.copy()) A__ = cmd_config[:-1] A__ = os.path.join(self.tmpdir , '''epoch_0''') cmd_config.extend( [ f"""--resume_from_checkpoint={resume_from_checkpoint}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(UpperCAmelCase__ , env=os.environ.copy()) def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->int: '''simple docstring''' A__ = os.path.join(self.test_scripts_folder , '''test_peak_memory_usage.py''') A__ = [ '''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): A__ = cmd.copy() if "fp16" in spec: cmd_config.extend(['''--mixed_precision=fp16''']) else: cmd_config.extend(['''--mixed_precision=no''']) if "multi_gpu" in spec: continue else: cmd_config.extend(['''--use_fsdp''']) for i, strategy in enumerate(UpperCAmelCase__): if strategy.lower() in spec: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""") break if "cpu_offload" in spec: cmd_config.append('''--fsdp_offload_params=True''') for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""") break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''') elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''') cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--peak_memory_upper_bound={peak_mem_upper_bound}""", f"""--n_train={self.n_train}""", f"""--n_val={self.n_val}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(UpperCAmelCase__ , env=os.environ.copy())
14
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase ) , """Tatoeba directory does not exist.""" ) class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Dict = self.resolver.write_model_card('opus-mt-he-en' ,dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
0
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE :List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Tuple = { 'transfo-xl-wt103': 'https://huggingface.co/transfo-xl-wt103/resolve/main/config.json', } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = "transfo-xl" snake_case_ = ["mems"] snake_case_ = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : int ,A : List[str]=26_77_35 ,A : Tuple=[2_00_00, 4_00_00, 20_00_00] ,A : int=10_24 ,A : Any=10_24 ,A : Dict=16 ,A : Union[str, Any]=64 ,A : Union[str, Any]=40_96 ,A : Union[str, Any]=4 ,A : Optional[Any]=False ,A : Any=18 ,A : Optional[int]=16_00 ,A : str=10_00 ,A : Optional[int]=True ,A : Dict=True ,A : Any=0 ,A : int=-1 ,A : str=True ,A : str=0.1 ,A : Dict=0.0 ,A : str=True ,A : Any="normal" ,A : List[str]=0.01 ,A : int=0.01 ,A : Optional[int]=0.02 ,A : int=1E-5 ,A : Optional[Any]=0 ,**A : Optional[int] ,): __A = vocab_size __A = [] self.cutoffs.extend(A ) if proj_share_all_but_first: __A = [False] + [True] * len(self.cutoffs ) else: __A = [False] + [False] * len(self.cutoffs ) __A = d_model __A = d_embed __A = d_head __A = d_inner __A = div_val __A = pre_lnorm __A = n_layer __A = n_head __A = mem_len __A = same_length __A = attn_type __A = clamp_len __A = sample_softmax __A = adaptive __A = dropout __A = dropatt __A = untie_r __A = init __A = init_range __A = proj_init_std __A = init_std __A = layer_norm_epsilon super().__init__(eos_token_id=A ,**A ) @property def UpperCamelCase_ ( self : List[str] ): # Message copied from Transformer-XL documentation logger.info(f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def UpperCamelCase_ ( self : Any ,A : Optional[Any] ): # Message copied from Transformer-XL documentation raise NotImplementedError( f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
15
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
0
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def __UpperCAmelCase ( __lowerCamelCase ) -> List[str]: lowercase__ : Optional[int] = FileLock(str(tmpdir / '''foo.lock''' ) ) lowercase__ : List[Any] = FileLock(str(tmpdir / '''foo.lock''' ) ) lowercase__ : int = 0.0_1 with locka.acquire(): with pytest.raises(__lowerCamelCase ): lowercase__ : int = time.time() locka.acquire(__lowerCamelCase ) assert time.time() - _start > timeout def __UpperCAmelCase ( __lowerCamelCase ) -> List[Any]: lowercase__ : Any = '''a''' * 10_00 + '''.lock''' lowercase__ : str = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__lowerCamelCase ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 lowercase__ : Dict = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__lowerCamelCase ): locka.acquire(0 )
16
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : Dict = 3 _UpperCamelCase : Any = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 1_5 == 0: result -= a a += 1 return result if __name__ == "__main__": print(F"""{solution() = }""")
83
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Union[str, Any] ): __lowercase = tempfile.mkdtemp() __lowercase = BlipImageProcessor() __lowercase = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-BertModel" ) __lowercase = BlipProcessor(UpperCAmelCase__, UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple, **UpperCAmelCase__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ).tokenizer def _lowercase ( self : List[str], **UpperCAmelCase__ : Tuple ): return AutoProcessor.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ).image_processor def _lowercase ( self : Dict ): shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ): __lowercase = [np.random.randint(2_5_5, size=(3, 3_0, 4_0_0), dtype=np.uinta )] __lowercase = [Image.fromarray(np.moveaxis(UpperCAmelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : List[Any] ): __lowercase = BlipProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __lowercase = self.get_tokenizer(bos_token="(BOS)", eos_token="(EOS)" ) __lowercase = self.get_image_processor(do_normalize=UpperCAmelCase__, padding_value=1.0 ) __lowercase = BlipProcessor.from_pretrained( self.tmpdirname, bos_token="(BOS)", eos_token="(EOS)", do_normalize=UpperCAmelCase__, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, UpperCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCAmelCase__ ) def _lowercase ( self : Dict ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = BlipProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = self.prepare_image_inputs() __lowercase = image_processor(UpperCAmelCase__, return_tensors="np" ) __lowercase = processor(images=UpperCAmelCase__, return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1E-2 ) def _lowercase ( self : Optional[int] ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = BlipProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = "lower newer" __lowercase = processor(text=UpperCAmelCase__ ) __lowercase = tokenizer(UpperCAmelCase__, return_token_type_ids=UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def _lowercase ( self : Tuple ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = BlipProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = "lower newer" __lowercase = self.prepare_image_inputs() __lowercase = processor(text=UpperCAmelCase__, images=UpperCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), ["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase__ ): processor() def _lowercase ( self : List[str] ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = BlipProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowercase = processor.batch_decode(UpperCAmelCase__ ) __lowercase = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__, UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = BlipProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = "lower newer" __lowercase = self.prepare_image_inputs() __lowercase = processor(text=UpperCAmelCase__, images=UpperCAmelCase__ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ), ["pixel_values", "input_ids", "attention_mask"] )
17
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
0
def _snake_case ( lowerCAmelCase : int ): """simple docstring""" return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('''Program to check whether a number is a Perfect number or not...''') __lowerCamelCase : Any = int(input('''Enter number: ''').strip()) print(f'''{number} is {'' if perfect(number) else 'not '}a Perfect Number.''')
18
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') snake_case_ : Any = logging.getLogger(__name__) @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) lowercase__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class lowercase__ : lowercase__ = field(default=lowercase , metadata={"""help""": """The input training data file (a text file)."""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def UpperCamelCase_ ( self : str ): '''simple docstring''' if self.train_file is not None: _UpperCamelCase : List[Any] = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _UpperCamelCase : Union[str, Any] = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowercase__ : lowercase__ = 42 lowercase__ = True lowercase__ = None lowercase__ = None def __call__( self : Optional[Any] ,lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = 'label' if 'label' in features[0].keys() else 'labels' _UpperCamelCase : List[Any] = [feature.pop(lowerCamelCase__ ) for feature in features] _UpperCamelCase : Dict = len(lowerCamelCase__ ) _UpperCamelCase : List[str] = len(features[0]['input_ids'] ) _UpperCamelCase : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(lowerCamelCase__ )] for feature in features ] _UpperCamelCase : str = list(chain(*lowerCamelCase__ ) ) _UpperCamelCase : Tuple = self.tokenizer.pad( lowerCamelCase__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='pt' ,) # Un-flatten _UpperCamelCase : str = {k: v.view(lowerCamelCase__ ,lowerCamelCase__ ,-1 ) for k, v in batch.items()} # Add back labels _UpperCamelCase : Optional[int] = torch.tensor(lowerCamelCase__ ,dtype=torch.intaa ) return batch def A__ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCamelCase : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Optional[int] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , UpperCAmelCase_ , UpperCAmelCase_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(UpperCAmelCase_ ) datasets.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCamelCase : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCamelCase : List[str] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _UpperCamelCase : Optional[int] = {} if data_args.train_file is not None: _UpperCamelCase : Tuple = data_args.train_file if data_args.validation_file is not None: _UpperCamelCase : Tuple = data_args.validation_file _UpperCamelCase : Any = data_args.train_file.split('.' )[-1] _UpperCamelCase : Union[str, Any] = load_dataset( UpperCAmelCase_ , data_files=UpperCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _UpperCamelCase : List[str] = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=UpperCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _UpperCamelCase : Any = [f'ending{i}' for i in range(4 )] _UpperCamelCase : int = 'sent1' _UpperCamelCase : List[str] = 'sent2' if data_args.max_seq_length is None: _UpperCamelCase : int = tokenizer.model_max_length if max_seq_length > 1_0_2_4: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) _UpperCamelCase : int = 1_0_2_4 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) _UpperCamelCase : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(UpperCAmelCase_ ): _UpperCamelCase : str = [[context] * 4 for context in examples[context_name]] _UpperCamelCase : Optional[Any] = examples[question_header_name] _UpperCamelCase : Tuple = [ [f'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(UpperCAmelCase_ ) ] # Flatten out _UpperCamelCase : Optional[int] = list(chain(*UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = list(chain(*UpperCAmelCase_ ) ) # Tokenize _UpperCamelCase : Tuple = tokenizer( UpperCAmelCase_ , UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(UpperCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _UpperCamelCase : Optional[Any] = raw_datasets['train'] if data_args.max_train_samples is not None: _UpperCamelCase : Tuple = min(len(UpperCAmelCase_ ) , data_args.max_train_samples ) _UpperCamelCase : Tuple = train_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): _UpperCamelCase : Union[str, Any] = train_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _UpperCamelCase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _UpperCamelCase : Union[str, Any] = min(len(UpperCAmelCase_ ) , data_args.max_eval_samples ) _UpperCamelCase : str = eval_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): _UpperCamelCase : Dict = eval_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _UpperCamelCase : List[Any] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=UpperCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Union[str, Any] = eval_predictions _UpperCamelCase : List[str] = np.argmax(UpperCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer _UpperCamelCase : Optional[int] = Trainer( model=UpperCAmelCase_ , args=UpperCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=UpperCAmelCase_ , data_collator=UpperCAmelCase_ , compute_metrics=UpperCAmelCase_ , ) # Training if training_args.do_train: _UpperCamelCase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _UpperCamelCase : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCamelCase : int = last_checkpoint _UpperCamelCase : List[str] = trainer.train(resume_from_checkpoint=UpperCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload _UpperCamelCase : Union[str, Any] = train_result.metrics _UpperCamelCase : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('train' , UpperCAmelCase_ ) trainer.save_metrics('train' , UpperCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _UpperCamelCase : List[Any] = trainer.evaluate() _UpperCamelCase : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(UpperCAmelCase_ ) _UpperCamelCase : int = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('eval' , UpperCAmelCase_ ) trainer.save_metrics('eval' , UpperCAmelCase_ ) _UpperCamelCase : Optional[int] = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**UpperCAmelCase_ ) else: trainer.create_model_card(**UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
83
0
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def lowerCamelCase_ ( lowerCamelCase__ ): return {key.lstrip("-" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def lowerCamelCase_ ( ): lowerCamelCase_ = ArgumentParser( "HuggingFace Datasets CLI tool" , usage="datasets-cli <command> [<args>]" , allow_abbrev=lowerCamelCase__ ) lowerCamelCase_ = parser.add_subparsers(help="datasets-cli command helpers" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(lowerCamelCase__ ) EnvironmentCommand.register_subcommand(lowerCamelCase__ ) TestCommand.register_subcommand(lowerCamelCase__ ) RunBeamCommand.register_subcommand(lowerCamelCase__ ) DummyDataCommand.register_subcommand(lowerCamelCase__ ) # Parse args lowerCamelCase_ , lowerCamelCase_ = parser.parse_known_args() if not hasattr(lowerCamelCase__ , "func" ): parser.print_help() exit(1 ) lowerCamelCase_ = parse_unknown_args(lowerCamelCase__ ) # Run lowerCamelCase_ = args.func(lowerCamelCase__ , **lowerCamelCase__ ) service.run() if __name__ == "__main__": main()
19
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) lowercase__ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) lowercase__ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def A__ ( ): _UpperCamelCase : Optional[Any] = HfArgumentParser((ModelArguments,) ) ((_UpperCamelCase) , ) : Optional[int] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _UpperCamelCase : Any = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _UpperCamelCase : List[Any] = True _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : str = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=UpperCAmelCase_ , decoder_config=UpperCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _UpperCamelCase : str = decoder_config.decoder_start_token_id _UpperCamelCase : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _UpperCamelCase : int = decoder_config.bos_token_id if pad_token_id is None: _UpperCamelCase : Dict = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _UpperCamelCase : List[Any] = decoder_config.eos_token_id _UpperCamelCase : Dict = decoder_start_token_id _UpperCamelCase : int = pad_token_id _UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
83
0
def _snake_case( SCREAMING_SNAKE_CASE__ ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("""The given input must be positive""" ) # get the generated string sequence lowercase : List[str] = gray_code_sequence_string(SCREAMING_SNAKE_CASE__ ) # # convert them to integers for i in range(len(SCREAMING_SNAKE_CASE__ ) ): lowercase : List[Any] = int(sequence[i] , 2 ) return sequence def _snake_case( SCREAMING_SNAKE_CASE__ ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowercase : Union[str, Any] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowercase : Optional[int] = gray_code_sequence_string(bit_count - 1 ) lowercase : Tuple = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowercase : Any = """0""" + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE__ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowercase : Dict = """1""" + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE__ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
20
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy snake_case_ : Dict = logging.get_logger(__name__) class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : float ,**lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : List[Any] = feature_size _UpperCamelCase : Any = sampling_rate _UpperCamelCase : Optional[Any] = padding_value _UpperCamelCase : Union[str, Any] = kwargs.pop('padding_side' ,'right' ) _UpperCamelCase : Dict = kwargs.pop('return_attention_mask' ,lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = True ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,): '''simple docstring''' # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowerCamelCase__ ,(list, tuple) ) and isinstance(processed_features[0] ,(dict, BatchFeature) ): _UpperCamelCase : int = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' F' to this method that includes {self.model_input_names[0]}, but you provided' F' {list(processed_features.keys() )}' ) _UpperCamelCase : List[Any] = processed_features[self.model_input_names[0]] _UpperCamelCase : Dict = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowerCamelCase__ ) == 0: if return_attention_mask: _UpperCamelCase : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch _UpperCamelCase : List[str] = required_input[0] if isinstance(lowerCamelCase__ ,(list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. _UpperCamelCase : List[str] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowerCamelCase__ ): _UpperCamelCase : Dict = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowerCamelCase__ ): _UpperCamelCase : Any = 'tf' elif is_torch_tensor(lowerCamelCase__ ): _UpperCamelCase : Optional[int] = 'pt' elif isinstance(lowerCamelCase__ ,(int, float, list, tuple, np.ndarray) ): _UpperCamelCase : int = 'np' else: raise ValueError( F'type of {first_element} unknown: {type(lowerCamelCase__ )}. ' 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] ,(int, float) ): _UpperCamelCase : Any = to_numpy(lowerCamelCase__ ) else: _UpperCamelCase : Any = [to_numpy(lowerCamelCase__ ) for v in value] # Convert padding_strategy in PaddingStrategy _UpperCamelCase : Optional[int] = self._get_padding_strategies(padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ) _UpperCamelCase : str = processed_features[self.model_input_names[0]] _UpperCamelCase : List[str] = len(lowerCamelCase__ ) if not all(len(lowerCamelCase__ ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) _UpperCamelCase : List[str] = [] for i in range(lowerCamelCase__ ): _UpperCamelCase : List[str] = {k: v[i] for k, v in processed_features.items()} # truncation _UpperCamelCase : List[str] = self._truncate( lowerCamelCase__ ,max_length=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,truncation=lowerCamelCase__ ,) truncated_inputs.append(lowerCamelCase__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length _UpperCamelCase : Union[str, Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) _UpperCamelCase : Any = PaddingStrategy.MAX_LENGTH _UpperCamelCase : Optional[Any] = {} for i in range(lowerCamelCase__ ): # padding _UpperCamelCase : Any = self._pad( truncated_inputs[i] ,max_length=lowerCamelCase__ ,padding_strategy=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,) for key, value in outputs.items(): if key not in batch_outputs: _UpperCamelCase : Dict = [] if value.dtype is np.dtype(np.floataa ): _UpperCamelCase : Any = value.astype(np.floataa ) batch_outputs[key].append(lowerCamelCase__ ) return BatchFeature(lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' _UpperCamelCase : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: _UpperCamelCase : Optional[Any] = len(lowerCamelCase__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : str = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : str = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowerCamelCase__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: _UpperCamelCase : Tuple = np.ones(len(lowerCamelCase__ ) ,dtype=np.intaa ) if needs_to_be_padded: _UpperCamelCase : Dict = max_length - len(lowerCamelCase__ ) if self.padding_side == "right": if return_attention_mask: _UpperCamelCase : Optional[int] = np.pad( processed_features['attention_mask'] ,(0, difference) ) _UpperCamelCase : Union[str, Any] = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) _UpperCamelCase : List[Any] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: _UpperCamelCase : List[Any] = np.pad( processed_features['attention_mask'] ,(difference, 0) ) _UpperCamelCase : List[Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) _UpperCamelCase : List[str] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) _UpperCamelCase : int = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : Optional[int] = len(lowerCamelCase__ ) > max_length if needs_to_be_truncated: _UpperCamelCase : Dict = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: _UpperCamelCase : Optional[Any] = processed_features['attention_mask'][:max_length] return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : int=False ,lowerCamelCase__ : Optional[Any]=None ): '''simple docstring''' # Get padding strategy if padding is not False: if padding is True: _UpperCamelCase : Optional[Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Tuple = PaddingStrategy(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = padding else: _UpperCamelCase : List[Any] = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
83
0
from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def UpperCamelCase_( lowerCamelCase_ ) -> Any: if isinstance(lowerCamelCase_ , collections.abc.Iterable ): return x return (x, x) @require_tf class _lowerCamelCase: def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> int: """simple docstring""" pass def UpperCamelCase ( self) -> str: """simple docstring""" pass def UpperCamelCase ( self) -> List[Any]: """simple docstring""" pass def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" _lowercase : str = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase, lowerCamelCase) _lowercase : Dict = TFVisionTextDualEncoderModel(lowerCamelCase) _lowercase : Dict = model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase) self.assertEqual(output['text_embeds'].shape, (input_ids.shape[0], config.projection_dim)) self.assertEqual(output['image_embeds'].shape, (pixel_values.shape[0], config.projection_dim)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> str: """simple docstring""" _lowercase , _lowercase : Optional[int] = self.get_vision_text_model(lowerCamelCase, lowerCamelCase) _lowercase : Optional[int] = TFVisionTextDualEncoderModel(vision_model=lowerCamelCase, text_model=lowerCamelCase) _lowercase : Dict = model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase) self.assertEqual(output['text_embeds'].shape, (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output['image_embeds'].shape, (pixel_values.shape[0], model.config.projection_dim)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Tuple: """simple docstring""" _lowercase , _lowercase : Optional[int] = self.get_vision_text_model(lowerCamelCase, lowerCamelCase) _lowercase : Any = {'vision_model': vision_model, 'text_model': text_model} _lowercase : Optional[int] = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase) _lowercase : Union[str, Any] = model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase) self.assertEqual(output['text_embeds'].shape, (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output['image_embeds'].shape, (pixel_values.shape[0], model.config.projection_dim)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" _lowercase , _lowercase : Tuple = self.get_vision_text_model(lowerCamelCase, lowerCamelCase) _lowercase : List[str] = TFVisionTextDualEncoderModel(vision_model=lowerCamelCase, text_model=lowerCamelCase) _lowercase : Union[str, Any] = model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase) _lowercase : Optional[int] = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase) _lowercase : Optional[int] = TFVisionTextDualEncoderModel.from_pretrained(lowerCamelCase) _lowercase : str = model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase) _lowercase : List[Any] = after_output[0].numpy() _lowercase : str = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(lowerCamelCase, 1E-5) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Dict: """simple docstring""" _lowercase , _lowercase : Union[str, Any] = self.get_vision_text_model(lowerCamelCase, lowerCamelCase) _lowercase : List[Any] = TFVisionTextDualEncoderModel(vision_model=lowerCamelCase, text_model=lowerCamelCase) _lowercase : int = model( input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, output_attentions=lowerCamelCase) _lowercase : Dict = output.vision_model_output.attentions self.assertEqual(len(lowerCamelCase), vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _lowercase : Optional[Any] = to_atuple(vision_model.config.image_size) _lowercase : Optional[int] = to_atuple(vision_model.config.patch_size) _lowercase : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _lowercase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:], (vision_config.num_attention_heads, seq_len, seq_len)) _lowercase : Any = output.text_model_output.attentions self.assertEqual(len(lowerCamelCase), text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:], (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]), ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" _lowercase : Optional[Any] = np.abs((a - b)).max() self.assertLessEqual(lowerCamelCase, lowerCamelCase, F'''Difference between torch and flax is {diff} (>= {tol}).''') def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : int = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : str = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Union[str, Any] = self.prepare_config_and_inputs() self.check_save_load(**lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : str = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowerCamelCase) @slow def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase , _lowercase : Tuple = self.get_pretrained_model_and_inputs() _lowercase : Optional[int] = model_a(**lowerCamelCase) _lowercase : Optional[int] = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowerCamelCase) _lowercase : Optional[Any] = TFVisionTextDualEncoderModel.from_pretrained(lowerCamelCase) _lowercase : Optional[int] = model_a(**lowerCamelCase) _lowercase : Optional[int] = after_outputs[0].numpy() _lowercase : str = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(lowerCamelCase, 1E-5) @require_tf class _lowerCamelCase( _a, unittest.TestCase ): def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : int = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-vit', 'hf-internal-testing/tiny-random-bert') _lowercase : Union[str, Any] = 13 _lowercase : Union[str, Any] = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) _lowercase : str = ids_tensor([batch_size, 4], model.text_model.config.vocab_size) _lowercase : Any = random_attention_mask([batch_size, 4]) _lowercase : List[Any] = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : int = TFViTModel(lowerCamelCase, name='vision_model') _lowercase : Union[str, Any] = TFBertModel(lowerCamelCase, name='text_model') return vision_model, text_model def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : int = TFViTModelTester(self) _lowercase : List[Any] = TFBertModelTester(self) _lowercase : Any = vit_model_tester.prepare_config_and_inputs() _lowercase : str = bert_model_tester.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase : str = vision_config_and_inputs ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Union[str, Any] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class _lowerCamelCase( _a, unittest.TestCase ): def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : str = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-deit-tf', 'hf-internal-testing/tiny-random-roberta') _lowercase : Dict = 13 _lowercase : Any = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) _lowercase : int = ids_tensor([batch_size, 4], model.text_model.config.vocab_size) _lowercase : Tuple = random_attention_mask([batch_size, 4]) _lowercase : Optional[Any] = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" _lowercase , _lowercase : Dict = self.get_vision_text_model(lowerCamelCase, lowerCamelCase) _lowercase : Optional[Any] = TFVisionTextDualEncoderModel(vision_model=lowerCamelCase, text_model=lowerCamelCase) _lowercase : List[Any] = model( input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, output_attentions=lowerCamelCase) _lowercase : List[Any] = output.vision_model_output.attentions self.assertEqual(len(lowerCamelCase), vision_config.num_hidden_layers) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowercase : Dict = to_atuple(vision_model.config.image_size) _lowercase : List[str] = to_atuple(vision_model.config.patch_size) _lowercase : List[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _lowercase : Any = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:], (vision_config.num_attention_heads, seq_len, seq_len)) _lowercase : Optional[Any] = output.text_model_output.attentions self.assertEqual(len(lowerCamelCase), text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:], (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]), ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" _lowercase : Tuple = TFDeiTModel(lowerCamelCase, name='vision_model') _lowercase : str = TFRobertaModel(lowerCamelCase, name='text_model') return vision_model, text_model def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Union[str, Any] = TFDeiTModelTester(self) _lowercase : Tuple = TFRobertaModelTester(self) _lowercase : Any = vit_model_tester.prepare_config_and_inputs() _lowercase : List[Any] = bert_model_tester.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase : Optional[Any] = vision_config_and_inputs ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Any = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class _lowerCamelCase( _a, unittest.TestCase ): def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-clip-tf', 'hf-internal-testing/tiny-random-bert') _lowercase : List[Any] = 13 _lowercase : int = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) _lowercase : List[str] = ids_tensor([batch_size, 4], model.text_model.config.vocab_size) _lowercase : List[Any] = random_attention_mask([batch_size, 4]) _lowercase : int = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : str = TFCLIPVisionModel(lowerCamelCase, name='vision_model') _lowercase : List[str] = TFBertModel(lowerCamelCase, name='text_model') return vision_model, text_model def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Any = TFCLIPVisionModelTester(self) _lowercase : Tuple = TFBertModelTester(self) _lowercase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() _lowercase : Any = bert_model_tester.prepare_config_and_inputs() _lowercase , _lowercase : Any = vision_config_and_inputs ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[int] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = TFVisionTextDualEncoderModel.from_pretrained( 'clip-italian/clip-italian', logit_scale_init_value=1.0, from_pt=lowerCamelCase) _lowercase : str = VisionTextDualEncoderProcessor.from_pretrained('clip-italian/clip-italian') _lowercase : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png') _lowercase : int = processor( text=['una foto di un gatto', 'una foto di un cane'], images=lowerCamelCase, padding=lowerCamelCase, return_tensors='np') _lowercase : Union[str, Any] = model(**lowerCamelCase) # verify the logits self.assertEqual(outputs.logits_per_image.shape, (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape, (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]), ) _lowercase : Any = np.array([[1.2_2_8_4_7_2_7, 0.3_1_0_4_1_2_2]]) self.assertTrue(np.allclose(outputs.logits_per_image.numpy(), lowerCamelCase, atol=1E-3))
21
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
0
'''simple docstring''' import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging __SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) class A_ ( lowerCAmelCase_ ): _lowerCamelCase : Optional[Any] = CLIPConfig _lowerCamelCase : Optional[Any] = ["""CLIPEncoderLayer"""] def __init__( self : Dict , snake_case_ : CLIPConfig ): super().__init__(snake_case_ ) _UpperCAmelCase = CLIPVisionModelWithProjection(config.vision_config ) _UpperCAmelCase = nn.Linear(config.vision_config.projection_dim , 1 ) _UpperCAmelCase = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def lowercase ( self : List[Any] , snake_case_ : Optional[int] , snake_case_ : Union[str, Any] , snake_case_ : Optional[int]=0.5 , snake_case_ : List[str]=0.5 ): _UpperCAmelCase = self.vision_model(snake_case_ )[0] _UpperCAmelCase = self.p_head(snake_case_ ) _UpperCAmelCase = nsfw_detected.flatten() _UpperCAmelCase = nsfw_detected > p_threshold _UpperCAmelCase = nsfw_detected.tolist() if any(snake_case_ ): logger.warning( "Potential NSFW content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, nsfw_detected_ in enumerate(snake_case_ ): if nsfw_detected_: _UpperCAmelCase = np.zeros(images[idx].shape ) _UpperCAmelCase = self.w_head(snake_case_ ) _UpperCAmelCase = watermark_detected.flatten() _UpperCAmelCase = watermark_detected > w_threshold _UpperCAmelCase = watermark_detected.tolist() if any(snake_case_ ): logger.warning( "Potential watermarked content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, watermark_detected_ in enumerate(snake_case_ ): if watermark_detected_: _UpperCAmelCase = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
22
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowercase__ ( lowercase ): @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Dict = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : Dict = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : str = '1' _UpperCamelCase : Union[str, Any] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : List[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : List[Any] = self.get_env() _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # next emulate no network _UpperCamelCase : Dict = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : Dict = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : int = '\nfrom transformers import pipeline\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _UpperCamelCase : Union[str, Any] = self.get_env() _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Tuple = [sys.executable, '-c', '\n'.join([load, mock, run] )] _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,1 ,result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' ,result.stderr.decode().replace('\n' ,'' ) ,) @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = '\nfrom transformers import AutoModel\n ' _UpperCamelCase : int = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Any = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : Optional[int] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() )
83
0
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class SCREAMING_SNAKE_CASE( datasets.BeamBasedBuilder ): """simple docstring""" def A ( self : Dict ) -> Tuple: return datasets.DatasetInfo( features=datasets.Features({'''content''': datasets.Value('''string''' )} ) , supervised_keys=__snake_case , ) def A ( self : Union[str, Any] , __snake_case : List[str] , __snake_case : str ) -> Optional[Any]: return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_dummy_examples()} )] def A ( self : Optional[Any] , __snake_case : Optional[Any] , __snake_case : Optional[Any] ) -> Optional[Any]: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(__snake_case ) class SCREAMING_SNAKE_CASE( datasets.BeamBasedBuilder ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: return datasets.DatasetInfo( features=datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) , supervised_keys=__snake_case , ) def A ( self : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] ) -> Tuple: return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_nested_examples()} ) ] def A ( self : Tuple , __snake_case : Tuple , __snake_case : str ) -> str: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(__snake_case ) def snake_case_ ( ) -> Optional[Any]: return [(i, {"content": content}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] def snake_case_ ( ) -> Union[str, Any]: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" @require_beam def A ( self : Tuple ) -> Dict: UpperCAmelCase : Dict = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase : str = DummyBeamDataset(cache_dir=__snake_case , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(__snake_case , builder.name , '''default''' , '''0.0.0''' , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCAmelCase : str = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __snake_case ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __snake_case ) self.assertDictEqual(dset['''train'''][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(__snake_case , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def A ( self : List[Any] ) -> List[Any]: import apache_beam as beam UpperCAmelCase : Tuple = beam.io.parquetio.WriteToParquet UpperCAmelCase : List[str] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase : Any = DummyBeamDataset(cache_dir=__snake_case , beam_runner='''DirectRunner''' ) with patch('''apache_beam.io.parquetio.WriteToParquet''' ) as write_parquet_mock: UpperCAmelCase : str = partial(__snake_case , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( __snake_case , builder.name , '''default''' , '''0.0.0''' , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertTrue( os.path.exists( os.path.join( __snake_case , builder.name , '''default''' , '''0.0.0''' , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) UpperCAmelCase : Any = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __snake_case ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __snake_case ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset['''train''']['''content'''] ) , sorted(['''foo''', '''bar''', '''foobar'''] ) ) self.assertTrue( os.path.exists(os.path.join(__snake_case , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def A ( self : Optional[Any] ) -> List[str]: with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase : Any = DummyBeamDataset(cache_dir=__snake_case ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def A ( self : Optional[Any] ) -> Optional[Any]: UpperCAmelCase : List[str] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase : List[str] = NestedBeamDataset(cache_dir=__snake_case , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(__snake_case , builder.name , '''default''' , '''0.0.0''' , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) ) UpperCAmelCase : List[Any] = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __snake_case ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __snake_case ) self.assertDictEqual(dset['''train'''][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(__snake_case , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset
23
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase__ ( unittest.TestCase ): def __init__( self : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=99 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=512 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : int=0.0_2 ,lowerCamelCase__ : int=4 ,): '''simple docstring''' _UpperCamelCase : List[Any] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Optional[Any] = is_training _UpperCamelCase : Optional[int] = use_attention_mask _UpperCamelCase : Any = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : str = intermediate_size _UpperCamelCase : int = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Optional[int] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_choices def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Union[str, Any] = None if self.use_attention_mask: _UpperCamelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=lowerCamelCase__ ,) return config, input_ids, attention_mask def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Optional[int] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase : Dict = model_class_name.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase : Tuple = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] _UpperCamelCase : Any = (1, 11, 768) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1E-4 ) )
83
0
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py snake_case_ = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. snake_case_ = direct_transformers_import(PATH_TO_TRANSFORMERS) snake_case_ = transformers.models.auto.configuration_auto.CONFIG_MAPPING snake_case_ = { # used to compute the property `self.chunk_length` 'EncodecConfig': ['overlap'], # used as `self.bert_model = BertModel(config, ...)` 'DPRConfig': True, # not used in modeling files, but it's an important information 'FSMTConfig': ['langs'], # used internally in the configuration class file 'GPTNeoConfig': ['attention_types'], # used internally in the configuration class file 'EsmConfig': ['is_folding_model'], # used during training (despite we don't have training script for these models yet) 'Mask2FormerConfig': ['ignore_value'], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) 'OneFormerConfig': ['ignore_value', 'norm'], # used during preprocessing and collation, see `collating_graphormer.py` 'GraphormerConfig': ['spatial_pos_max'], # used internally in the configuration class file 'T5Config': ['feed_forward_proj'], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally 'MT5Config': ['feed_forward_proj', 'tokenizer_class'], 'UMT5Config': ['feed_forward_proj', 'tokenizer_class'], # used internally in the configuration class file 'LongT5Config': ['feed_forward_proj'], # used internally in the configuration class file 'SwitchTransformersConfig': ['feed_forward_proj'], # having default values other than `1e-5` - we can't fix them without breaking 'BioGptConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'GLPNConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'SegformerConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'CvtConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'PerceiverConfig': ['layer_norm_eps'], # used internally to calculate the feature size 'InformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'TimeSeriesTransformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'AutoformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate `mlp_dim` 'SamVisionConfig': ['mlp_ratio'], # For (head) training, but so far not implemented 'ClapAudioConfig': ['num_classes'], # Not used, but providing useful information to users 'SpeechT5HifiGanConfig': ['sampling_rate'], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { 'CLIPSegConfig': True, 'DeformableDetrConfig': True, 'DetaConfig': True, 'DinatConfig': True, 'DonutSwinConfig': True, 'EfficientFormerConfig': True, 'FSMTConfig': True, 'JukeboxConfig': True, 'LayoutLMv2Config': True, 'MaskFormerSwinConfig': True, 'MT5Config': True, 'NatConfig': True, 'OneFormerConfig': True, 'PerceiverConfig': True, 'RagConfig': True, 'SpeechT5Config': True, 'SwinConfig': True, 'Swin2SRConfig': True, 'Swinv2Config': True, 'SwitchTransformersConfig': True, 'TableTransformerConfig': True, 'TapasConfig': True, 'TransfoXLConfig': True, 'UniSpeechConfig': True, 'UniSpeechSatConfig': True, 'WavLMConfig': True, 'WhisperConfig': True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) 'JukeboxPriorConfig': True, # TODO: @Younes (for `is_decoder`) 'Pix2StructTextConfig': True, } ) def lowerCamelCase__ ( snake_case_ : List[str] , snake_case_ : Tuple , snake_case_ : Any , snake_case_ : List[Any] ) -> Union[str, Any]: __snake_case = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( f"""config.{attribute}""" in modeling_source or f"""getattr(config, \"{attribute}\"""" in modeling_source or f"""getattr(self.config, \"{attribute}\"""" in modeling_source ): __snake_case = True # Deal with multi-line cases elif ( re.search( Rf"""getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*\"{attribute}\"""" , snake_case_ , ) is not None ): __snake_case = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: __snake_case = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files __snake_case = [ '''bos_index''', '''eos_index''', '''pad_index''', '''unk_index''', '''mask_index''', '''image_size''', '''use_cache''', '''out_features''', '''out_indices''', ] __snake_case = ['''encoder_no_repeat_ngram_size'''] # Special cases to be allowed __snake_case = True if not attribute_used: __snake_case = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: __snake_case = True elif attribute in ["tie_word_embeddings"] and default_value is False: __snake_case = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: __snake_case = True elif attribute.endswith('''_token_id''' ): __snake_case = True # configuration class specific cases if not case_allowed: __snake_case = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , [] ) __snake_case = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def lowerCamelCase__ ( snake_case_ : Union[str, Any] ) -> Tuple: __snake_case = dict(inspect.signature(config_class.__init__ ).parameters ) __snake_case = [x for x in list(signature.keys() ) if x not in ['''self''', '''kwargs''']] __snake_case = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass __snake_case = {} if len(config_class.attribute_map ) > 0: __snake_case = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files __snake_case = inspect.getsourcefile(snake_case_ ) __snake_case = os.path.dirname(snake_case_ ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. __snake_case = [os.path.join(snake_case_ , snake_case_ ) for fn in os.listdir(snake_case_ ) if fn.startswith('''modeling_''' )] # Get the source code strings __snake_case = [] for path in modeling_paths: if os.path.isfile(snake_case_ ): with open(snake_case_ ) as fp: modeling_sources.append(fp.read() ) __snake_case = [] for config_param, default_value in zip(snake_case_ , snake_case_ ): # `attributes` here is all the variant names for `config_param` __snake_case = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(snake_case_ , snake_case_ , snake_case_ , snake_case_ ): unused_attributes.append(attributes[0] ) return sorted(snake_case_ ) def lowerCamelCase__ ( ) -> Union[str, Any]: __snake_case = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) __snake_case = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ) , lambda snake_case_ : inspect.isclass(snake_case_ ) and issubclass(snake_case_ , snake_case_ ) and inspect.getmodule(snake_case_ ) == inspect.getmodule(_config_class ) , ) ] for config_class in config_classes_in_module: __snake_case = check_config_attributes_being_used(snake_case_ ) if len(snake_case_ ) > 0: __snake_case = unused_attributes if len(snake_case_ ) > 0: __snake_case = '''The following configuration classes contain unused attributes in the corresponding modeling files:\n''' for name, attributes in configs_with_unused_attributes.items(): error += f"""{name}: {attributes}\n""" raise ValueError(snake_case_ ) if __name__ == "__main__": check_config_attributes()
24
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer snake_case_ : List[Any] = logging.get_logger(__name__) class lowercase__ ( lowercase ): lowercase__ = """AutoTokenizer""" lowercase__ = ["""tokenizer"""] lowercase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self : List[str] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Tuple=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : Dict = speaker_embeddings @classmethod def UpperCamelCase_ ( cls : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : str="speaker_embeddings_path.json" ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: _UpperCamelCase : Optional[Any] = get_file_from_repo( lowerCamelCase__ ,lowerCamelCase__ ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(lowerCamelCase__ ,lowerCamelCase__ )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) _UpperCamelCase : Union[str, Any] = None else: with open(lowerCamelCase__ ) as speaker_embeddings_json: _UpperCamelCase : Optional[int] = json.load(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = None _UpperCamelCase : Tuple = AutoTokenizer.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) return cls(tokenizer=lowerCamelCase__ ,speaker_embeddings=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int="speaker_embeddings_path.json" ,lowerCamelCase__ : Dict="speaker_embeddings" ,lowerCamelCase__ : bool = False ,**lowerCamelCase__ : Tuple ,): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ,'v2' ) ,exist_ok=lowerCamelCase__ ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Optional[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCamelCase : Any = self._load_voice_preset(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] ,lowerCamelCase__ ,F'{prompt_key}_{key}' ) ,voice_preset[key] ,allow_pickle=lowerCamelCase__ ,) _UpperCamelCase : List[str] = os.path.join(lowerCamelCase__ ,F'{prompt_key}_{key}.npy' ) _UpperCamelCase : str = tmp_dict with open(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ) ,'w' ) as fp: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) super().save_pretrained(lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : str = None ,**lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.speaker_embeddings[voice_preset] _UpperCamelCase : Union[str, Any] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) _UpperCamelCase : Dict = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' ,'/' ) ,voice_preset_paths[key] ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" ,"/" ) ,voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) _UpperCamelCase : List[str] = np.load(lowerCamelCase__ ) return voice_preset_dict def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] ,np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self : Any ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any="pt" ,lowerCamelCase__ : Dict=256 ,lowerCamelCase__ : int=False ,lowerCamelCase__ : int=True ,lowerCamelCase__ : List[str]=False ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' if voice_preset is not None and not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if ( isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCamelCase : Optional[int] = self._load_voice_preset(lowerCamelCase__ ) else: if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and not voice_preset.endswith('.npz' ): _UpperCamelCase : Tuple = voice_preset + '.npz' _UpperCamelCase : str = np.load(lowerCamelCase__ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.tokenizer( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,padding='max_length' ,max_length=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,**lowerCamelCase__ ,) if voice_preset is not None: _UpperCamelCase : Optional[Any] = voice_preset return encoded_text
83
0
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase__ : Any = get_tests_dir('fixtures/test_sentencepiece_bpe_char.model') @require_sentencepiece @require_tokenizers class lowerCAmelCase_ (a__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Dict = SpeechTaTokenizer __UpperCamelCase : Optional[Any] = False __UpperCamelCase : Any = True def __magic_name__ (self ) -> Dict: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE__ : Optional[int] = SpeechTaTokenizer(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = AddedToken("""<mask>""" , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[str] = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = """this is a test""" SCREAMING_SNAKE_CASE__ : List[str] = """this is a test""" return input_text, output_text def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=20 , SCREAMING_SNAKE_CASE__=5 ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : str = self.get_input_output_texts(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Any = tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[str] = tokenizer.decode(SCREAMING_SNAKE_CASE__ , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE__ ) return text, ids def __magic_name__ (self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = """<pad>""" SCREAMING_SNAKE_CASE__ : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-4] , """œ""" ) self.assertEqual(vocab_keys[-2] , """<mask>""" ) self.assertEqual(vocab_keys[-1] , """<ctc_blank>""" ) self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , 81 ) def __magic_name__ (self ) -> List[Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def __magic_name__ (self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.get_tokenizers(do_lower_case=SCREAMING_SNAKE_CASE__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Dict = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : Tuple = len(SCREAMING_SNAKE_CASE__ ) self.assertNotEqual(SCREAMING_SNAKE_CASE__ , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) SCREAMING_SNAKE_CASE__ : Optional[int] = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] SCREAMING_SNAKE_CASE__ : Any = tokenizer.add_tokens(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : int = len(SCREAMING_SNAKE_CASE__ ) self.assertNotEqual(SCREAMING_SNAKE_CASE__ , 0 ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) self.assertEqual(SCREAMING_SNAKE_CASE__ , all_size + len(SCREAMING_SNAKE_CASE__ ) ) SCREAMING_SNAKE_CASE__ : Optional[Any] = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=SCREAMING_SNAKE_CASE__ ) self.assertGreaterEqual(len(SCREAMING_SNAKE_CASE__ ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) SCREAMING_SNAKE_CASE__ : List[Any] = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.add_special_tokens(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : str = len(SCREAMING_SNAKE_CASE__ ) self.assertNotEqual(SCREAMING_SNAKE_CASE__ , 0 ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) self.assertEqual(SCREAMING_SNAKE_CASE__ , all_size_a + len(SCREAMING_SNAKE_CASE__ ) ) SCREAMING_SNAKE_CASE__ : int = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=SCREAMING_SNAKE_CASE__ ) self.assertGreaterEqual(len(SCREAMING_SNAKE_CASE__ ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def __magic_name__ (self ) -> Dict: """simple docstring""" pass def __magic_name__ (self ) -> Tuple: """simple docstring""" pass def __magic_name__ (self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(SCREAMING_SNAKE_CASE__ , [SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) SCREAMING_SNAKE_CASE__ : str = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( SCREAMING_SNAKE_CASE__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) # fmt: off self.assertListEqual(SCREAMING_SNAKE_CASE__ , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on SCREAMING_SNAKE_CASE__ : Any = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__ ) self.assertListEqual( SCREAMING_SNAKE_CASE__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def __magic_name__ (self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off SCREAMING_SNAKE_CASE__ : Optional[Any] = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=SCREAMING_SNAKE_CASE__ , model_name="""microsoft/speecht5_asr""" , revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" , sequences=SCREAMING_SNAKE_CASE__ , )
25
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin snake_case_ : Tuple = random.Random() def A__ ( UpperCAmelCase_ , UpperCAmelCase_=1.0 , UpperCAmelCase_=None , UpperCAmelCase_=None ): if rng is None: _UpperCamelCase : Dict = global_rng _UpperCamelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int=7 ,lowerCamelCase__ : str=400 ,lowerCamelCase__ : int=2000 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=16000 ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Optional[int]=True ,): '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Union[str, Any] = batch_size _UpperCamelCase : List[str] = min_seq_length _UpperCamelCase : Optional[int] = max_seq_length _UpperCamelCase : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _UpperCamelCase : List[str] = feature_size _UpperCamelCase : List[str] = padding_value _UpperCamelCase : List[Any] = sampling_rate _UpperCamelCase : Dict = return_attention_mask _UpperCamelCase : Tuple = do_normalize def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : Tuple=False ): '''simple docstring''' def _flatten(lowerCamelCase__ : Optional[Any] ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _UpperCamelCase : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _UpperCamelCase : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: _UpperCamelCase : int = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = WavaVecaFeatureExtractor def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): '''simple docstring''' self.assertTrue(np.all(np.mean(lowerCamelCase__ ,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ,axis=0 ) - 1 ) < 1E-3 ) ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' # Tests that all call wrap to encode_plus and batch_encode_plus _UpperCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _UpperCamelCase : int = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Tuple = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test not batched input _UpperCamelCase : Tuple = feat_extract(speech_inputs[0] ,return_tensors='np' ).input_values _UpperCamelCase : Any = feat_extract(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test batched _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : Optional[int] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _UpperCamelCase : str = [floats_list((1, x) )[0] for x in (800, 800, 800)] _UpperCamelCase : str = np.asarray(lowerCamelCase__ ) _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : int = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : str = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : List[str] = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,return_tensors='np' ) _UpperCamelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[str] = range(800 ,1400 ,200 ) _UpperCamelCase : List[str] = [floats_list((1, x) )[0] for x in lengths] _UpperCamelCase : Optional[Any] = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : str = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,max_length=lowerCamelCase__ ,padding=lowerCamelCase__ ) _UpperCamelCase : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Union[str, Any] = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='max_length' ,return_tensors='np' ) _UpperCamelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : int = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Any = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=2000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase_ ( self : Any ): '''simple docstring''' import torch _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = np.random.rand(100 ).astype(np.floataa ) _UpperCamelCase : Dict = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _UpperCamelCase : Optional[int] = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _UpperCamelCase : Tuple = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: _UpperCamelCase : Optional[int] = WavaVecaConfig.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase__ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask ,config.feat_extract_norm == 'layer' )
83
0
from math import ceil, sqrt def lowerCAmelCase_ ( snake_case_ = 1000000 ): _A : int = 0 for outer_width in range(3,(limit // 4) + 2 ): if outer_width**2 > limit: _A : List[str] = max(ceil(sqrt(outer_width**2 - limit ) ),1 ) else: _A : Optional[int] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f"""{solution() = }""")
26
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1 , UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : int = 1 _UpperCamelCase : Union[str, Any] = 0 for divide_by_number in range(UpperCAmelCase_ , digit + 1 ): _UpperCamelCase : list[int] = [] _UpperCamelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = len(UpperCAmelCase_ ) _UpperCamelCase : List[Any] = divide_by_number else: has_been_divided.append(UpperCAmelCase_ ) _UpperCamelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
83
0
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( lowerCAmelCase_ ): def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[str] = tempfile.mkdtemp() __a : Dict = 8 # DPR tok __a : List[Any] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __a : int = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__a , exist_ok=__a ) __a : List[str] = os.path.join(__a , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __a : Tuple = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __a : List[str] = dict(zip(__a , range(len(__a ) ) ) ) __a : Dict = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __a : Any = {'unk_token': '<unk>'} __a : int = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__a , exist_ok=__a ) __a : str = os.path.join(__a , BART_VOCAB_FILES_NAMES['vocab_file'] ) __a : List[str] = os.path.join(__a , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__a ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__a ) ) def __UpperCAmelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __UpperCAmelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __UpperCAmelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def __UpperCAmelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[Any] = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = self.get_dummy_dataset() __a : Dict = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __a : Union[str, Any] = dataset __a : Dict = RagRetriever( __a , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __UpperCAmelCase ( self , __a ): '''simple docstring''' __a : Optional[int] = self.get_dummy_dataset() __a : Tuple = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __a : Any = os.path.join(self.tmpdirname , 'dataset' ) __a : Union[str, Any] = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __a : int = RagRetriever( __a , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __a : List[Any] = RagRetriever( __a , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __a ) , ) return retriever def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __a : Optional[Any] = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __a : Dict = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __a : Union[str, Any] = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__a , open(__a , 'wb' ) ) __a : Union[str, Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __a : Optional[Any] = RagRetriever( __a , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = 1 __a : Union[str, Any] = self.get_dummy_canonical_hf_index_retriever() __a : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a , __a , __a : Union[str, Any] = retriever.retrieve(__a , n_docs=__a ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__a ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __a ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __a : Union[str, Any] = self.get_dummy_dataset() retriever.save_pretrained(__a ) __a : List[str] = RagRetriever.from_pretrained(__a ) self.assertIsInstance(__a , __a ) __a : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : Optional[Any] = retriever.retrieve(__a , n_docs=1 ) self.assertTrue(out is not None ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = 1 __a : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=__a ) __a : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a , __a , __a : str = retriever.retrieve(__a , n_docs=__a ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__a ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __a ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.get_dummy_custom_hf_index_retriever(from_disk=__a ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__a ) __a : Dict = RagRetriever.from_pretrained(__a ) self.assertIsInstance(__a , __a ) __a : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : List[str] = retriever.retrieve(__a , n_docs=1 ) self.assertTrue(out is not None ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = 1 __a : List[str] = self.get_dummy_custom_hf_index_retriever(from_disk=__a ) __a : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a , __a , __a : Any = retriever.retrieve(__a , n_docs=__a ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__a ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __a ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = self.get_dummy_custom_hf_index_retriever(from_disk=__a ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__a ) __a : int = RagRetriever.from_pretrained(__a ) self.assertIsInstance(__a , __a ) __a : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : Optional[Any] = retriever.retrieve(__a , n_docs=1 ) self.assertTrue(out is not None ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = 1 __a : str = self.get_dummy_legacy_index_retriever() __a : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a , __a , __a : int = retriever.retrieve(__a , n_docs=__a ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__a ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __a ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : str = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__a ) __a : Any = RagRetriever.from_pretrained(__a ) self.assertIsInstance(__a , __a ) __a : List[str] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : Dict = retriever.retrieve(__a , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __UpperCAmelCase ( self ): '''simple docstring''' import torch __a : Optional[int] = 1 __a : Any = self.get_dummy_canonical_hf_index_retriever() __a : Optional[int] = [[5, 7], [10, 11]] __a : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : List[str] = retriever(__a , __a , prefix=retriever.config.generator.prefix , n_docs=__a ) __a , __a , __a : Dict = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__a , __a ) self.assertIsInstance(__a , __a ) self.assertIsInstance(__a , np.ndarray ) __a : int = retriever( __a , __a , prefix=retriever.config.generator.prefix , n_docs=__a , return_tensors='pt' , ) __a , __a , __a , __a : List[str] = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__a , torch.Tensor ) self.assertIsInstance(__a , torch.Tensor ) self.assertIsInstance(__a , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.get_dpr_ctx_encoder_tokenizer() __a : Optional[int] = 1 __a : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=__a ) retriever.set_ctx_encoder_tokenizer(__a ) __a : int = [[5, 7], [10, 11]] __a : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __a : Optional[Any] = retriever(__a , __a , prefix=retriever.config.generator.prefix , n_docs=__a ) self.assertEqual( len(__a ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __a ) # check for doc token related keys in dictionary.
27
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if num < 0: return False _UpperCamelCase : int = num _UpperCamelCase : int = 0 while num > 0: _UpperCamelCase : str = rev_num * 1_0 + (num % 1_0) num //= 1_0 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
83
0
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Tuple = logging.get_logger(__name__) _lowerCamelCase : Any = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def __lowerCamelCase ( A__ ) -> List[str]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: UpperCamelCase = k.replace(A__ , A__ ) if k.startswith('encoder' ): UpperCamelCase = k.replace('.attn' , '.self_attn' ) UpperCamelCase = k.replace('norm1' , 'self_attn_layer_norm' ) UpperCamelCase = k.replace('norm2' , 'final_layer_norm' ) elif k.startswith('decoder' ): UpperCamelCase = k.replace('norm1' , 'self_attn_layer_norm' ) UpperCamelCase = k.replace('norm2' , 'encoder_attn_layer_norm' ) UpperCamelCase = k.replace('norm3' , 'final_layer_norm' ) return k def __lowerCamelCase ( A__ ) -> Any: """simple docstring""" UpperCamelCase = [ 'model.encoder.layernorm_embedding.weight', 'model.encoder.layernorm_embedding.bias', 'model.decoder.layernorm_embedding.weight', 'model.decoder.layernorm_embedding.bias', ] for k in keys: UpperCamelCase = sd.pop(A__ ) UpperCamelCase = k.replace('layernorm_embedding' , 'layer_norm' ) assert new_k not in sd UpperCamelCase = v _lowerCamelCase : Any = ["START"] @torch.no_grad() def __lowerCamelCase ( A__ , A__ , A__ ) -> int: """simple docstring""" UpperCamelCase = torch.load(A__ , map_location='cpu' ) UpperCamelCase = model['model'] UpperCamelCase = BlenderbotConfig.from_json_file(A__ ) UpperCamelCase = BlenderbotForConditionalGeneration(A__ ) UpperCamelCase = m.model.state_dict().keys() UpperCamelCase = [] UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue UpperCamelCase = rename_state_dict_key(A__ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(A__ ) m.model.load_state_dict(A__ , strict=A__ ) m.half() m.save_pretrained(A__ ) if __name__ == "__main__": _lowerCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) _lowerCamelCase : str = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
28
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = abs(UpperCAmelCase_ ) _UpperCamelCase : int = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = abs(UpperCAmelCase_ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def A__ ( UpperCAmelCase_ ): return sum(int(UpperCAmelCase_ ) for c in str(abs(UpperCAmelCase_ ) ) ) def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) -> None: _UpperCamelCase : str = f'{func.__name__}({value})' _UpperCamelCase : Tuple = timeit(f'__main__.{call}' , setup='import __main__' ) print(f'{call:56} = {func(UpperCAmelCase_ )} -- {timing:.4f} seconds' ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
83
0
# A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def lowercase__ ( __snake_case : List[Any] ): '''simple docstring''' UpperCAmelCase_ : Any = [False] * len(__snake_case ) UpperCAmelCase_ : Dict = [-1] * len(__snake_case ) def dfs(__snake_case : Dict , __snake_case : Tuple ): UpperCAmelCase_ : Optional[Any] = True UpperCAmelCase_ : Dict = c for u in graph[v]: if not visited[u]: dfs(__snake_case , 1 - c ) for i in range(len(__snake_case ) ): if not visited[i]: dfs(__snake_case , 0 ) for i in range(len(__snake_case ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph __UpperCAmelCase = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
29
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
0
def a ( snake_case__: int ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise TypeError('''Input value must be an \'int\' type''' ) lowercase_ = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
30
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
0
'''simple docstring''' import torch from torch import nn class lowerCamelCase_ (nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , A : Dict , A : Tuple , A : Optional[Any] , A : Tuple , A : Union[str, Any]=1 , A : str=False ): super().__init__() _UpperCAmelCase : Union[str, Any] = n_token _UpperCAmelCase : List[Any] = d_embed _UpperCAmelCase : List[str] = d_proj _UpperCAmelCase : Union[str, Any] = cutoffs + [n_token] _UpperCAmelCase : str = [0] + self.cutoffs _UpperCAmelCase : Dict = div_val _UpperCAmelCase : Tuple = self.cutoffs[0] _UpperCAmelCase : Tuple = len(self.cutoffs ) - 1 _UpperCAmelCase : Tuple = self.shortlist_size + self.n_clusters if self.n_clusters > 0: _UpperCAmelCase : Any = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) _UpperCAmelCase : Optional[int] = nn.Parameter(torch.zeros(self.n_clusters ) ) _UpperCAmelCase : str = nn.ModuleList() _UpperCAmelCase : Dict = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(A , A ) ) ) else: self.out_projs.append(A ) self.out_layers.append(nn.Linear(A , A ) ) else: for i in range(len(self.cutoffs ) ): _UpperCAmelCase , _UpperCAmelCase : Dict = self.cutoff_ends[i], self.cutoff_ends[i + 1] _UpperCAmelCase : Tuple = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(A , A ) ) ) self.out_layers.append(nn.Linear(A , r_idx - l_idx ) ) _UpperCAmelCase : Dict = keep_order def _A ( self : Optional[int] , A : Optional[int] , A : List[str] , A : Optional[Any] , A : List[str] ): if proj is None: _UpperCAmelCase : Optional[int] = nn.functional.linear(A , A , bias=A ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: _UpperCAmelCase : Optional[int] = nn.functional.linear(A , proj.t().contiguous() ) _UpperCAmelCase : str = nn.functional.linear(A , A , bias=A ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def _A ( self : Optional[int] , A : Tuple , A : Union[str, Any]=None , A : Any=False ): if labels is not None: # Shift so that tokens < n predict n _UpperCAmelCase : Union[str, Any] = hidden[..., :-1, :].contiguous() _UpperCAmelCase : str = labels[..., 1:].contiguous() _UpperCAmelCase : Any = hidden.view(-1 , hidden.size(-1 ) ) _UpperCAmelCase : str = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError("Input and labels should have the same size in the batch dimension." ) else: _UpperCAmelCase : Union[str, Any] = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: _UpperCAmelCase : Any = self._compute_logit(A , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: _UpperCAmelCase : Optional[int] = labels != -100 _UpperCAmelCase : List[str] = torch.zeros_like(A , dtype=hidden.dtype , device=hidden.device ) _UpperCAmelCase : List[str] = ( -nn.functional.log_softmax(A , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: _UpperCAmelCase : str = nn.functional.log_softmax(A , dim=-1 ) else: # construct weights and biases _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: _UpperCAmelCase , _UpperCAmelCase : List[Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _UpperCAmelCase : Any = self.out_layers[0].weight[l_idx:r_idx] _UpperCAmelCase : Union[str, Any] = self.out_layers[0].bias[l_idx:r_idx] else: _UpperCAmelCase : Optional[int] = self.out_layers[i].weight _UpperCAmelCase : Dict = self.out_layers[i].bias if i == 0: _UpperCAmelCase : int = torch.cat([weight_i, self.cluster_weight] , dim=0 ) _UpperCAmelCase : Union[str, Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(A ) biases.append(A ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[str] = weights[0], biases[0], self.out_projs[0] _UpperCAmelCase : Dict = self._compute_logit(A , A , A , A ) _UpperCAmelCase : List[str] = nn.functional.log_softmax(A , dim=1 ) if labels is None: _UpperCAmelCase : int = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: _UpperCAmelCase : int = torch.zeros_like(A , dtype=hidden.dtype , device=hidden.device ) _UpperCAmelCase : Dict = 0 _UpperCAmelCase : Any = [0] + self.cutoffs for i in range(len(A ) - 1 ): _UpperCAmelCase , _UpperCAmelCase : str = cutoff_values[i], cutoff_values[i + 1] if labels is not None: _UpperCAmelCase : List[Any] = (labels >= l_idx) & (labels < r_idx) _UpperCAmelCase : Any = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue _UpperCAmelCase : Union[str, Any] = labels.index_select(0 , A ) - l_idx _UpperCAmelCase : Any = head_logprob.index_select(0 , A ) _UpperCAmelCase : Dict = hidden.index_select(0 , A ) else: _UpperCAmelCase : List[Any] = hidden if i == 0: if labels is not None: _UpperCAmelCase : str = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: _UpperCAmelCase : Optional[Any] = head_logprob[:, : self.cutoffs[0]] else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = weights[i], biases[i], self.out_projs[i] _UpperCAmelCase : Optional[Any] = self._compute_logit(A , A , A , A ) _UpperCAmelCase : Optional[int] = nn.functional.log_softmax(A , dim=1 ) _UpperCAmelCase : Tuple = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: _UpperCAmelCase : Union[str, Any] = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: _UpperCAmelCase : str = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i _UpperCAmelCase : Optional[Any] = logprob_i if labels is not None: if (hasattr(self , "keep_order" ) and self.keep_order) or keep_order: out.index_copy_(0 , A , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def _A ( self : Optional[int] , A : str ): if self.n_clusters == 0: _UpperCAmelCase : List[str] = self._compute_logit(A , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(A , dim=-1 ) else: # construct weights and biases _UpperCAmelCase , _UpperCAmelCase : List[Any] = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _UpperCAmelCase : Union[str, Any] = self.out_layers[0].weight[l_idx:r_idx] _UpperCAmelCase : List[Any] = self.out_layers[0].bias[l_idx:r_idx] else: _UpperCAmelCase : int = self.out_layers[i].weight _UpperCAmelCase : List[str] = self.out_layers[i].bias if i == 0: _UpperCAmelCase : Tuple = torch.cat([weight_i, self.cluster_weight] , dim=0 ) _UpperCAmelCase : Any = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(A ) biases.append(A ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Optional[int] = weights[0], biases[0], self.out_projs[0] _UpperCAmelCase : Optional[Any] = self._compute_logit(A , A , A , A ) _UpperCAmelCase : Union[str, Any] = hidden.new_empty((head_logit.size(0 ), self.n_token) ) _UpperCAmelCase : Any = nn.functional.log_softmax(A , dim=1 ) _UpperCAmelCase : Optional[Any] = [0] + self.cutoffs for i in range(len(A ) - 1 ): _UpperCAmelCase , _UpperCAmelCase : List[str] = cutoff_values[i], cutoff_values[i + 1] if i == 0: _UpperCAmelCase : str = head_logprob[:, : self.cutoffs[0]] else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Tuple = weights[i], biases[i], self.out_projs[i] _UpperCAmelCase : int = self._compute_logit(A , A , A , A ) _UpperCAmelCase : List[str] = nn.functional.log_softmax(A , dim=1 ) _UpperCAmelCase : Optional[Any] = head_logprob[:, -i] + tail_logprob_i _UpperCAmelCase : Any = logprob_i return out
31
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
0
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness UpperCAmelCase_ : List[str] = '\\n@misc{chen2021evaluating,\n title={Evaluating Large Language Models Trained on Code},\n author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \\nand Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \\nand Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \\nand Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \\nand Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \\nand Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \\nand Mohammad Bavarian and Clemens Winter and Philippe Tillet \\nand Felipe Petroski Such and Dave Cummings and Matthias Plappert \\nand Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \\nand William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \\nand Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \\nand William Saunders and Christopher Hesse and Andrew N. Carr \\nand Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \\nand Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \\nand Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \\nand Sam McCandlish and Ilya Sutskever and Wojciech Zaremba},\n year={2021},\n eprint={2107.03374},\n archivePrefix={arXiv},\n primaryClass={cs.LG}\n}\n' UpperCAmelCase_ : Any = '\\nThis metric implements the evaluation harness for the HumanEval problem solving dataset\ndescribed in the paper "Evaluating Large Language Models Trained on Code"\n(https://arxiv.org/abs/2107.03374).\n' UpperCAmelCase_ : List[str] = '\nCalculates how good are predictions given some references, using certain scores\nArgs:\n predictions: list of candidates to evaluate. Each candidates should be a list\n of strings with several code candidates to solve the problem.\n references: a list with a test for each prediction. Each test should evaluate the\n correctness of a code candidate.\n k: number of code candidates to consider in the evaluation (Default: [1, 10, 100])\n num_workers: number of workers used to evaluate the canidate programs (Default: 4).\n timeout:\nReturns:\n pass_at_k: dict with pass rates for each k\n results: dict with granular results of each unittest\nExamples:\n >>> code_eval = datasets.load_metric("code_eval")\n >>> test_cases = ["assert add(2,3)==5"]\n >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]]\n >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2])\n >>> print(pass_at_k)\n {\'pass@1\': 0.5, \'pass@2\': 1.0}\n' UpperCAmelCase_ : Any = '\n################################################################################\n !!!WARNING!!!\n################################################################################\nThe "code_eval" metric executes untrusted model-generated code in Python.\nAlthough it is highly unlikely that model-generated code will do something\novertly malicious in response to this test suite, model-generated code may act\ndestructively due to a lack of model capability or alignment.\nUsers are strongly encouraged to sandbox this evaluation suite so that it\ndoes not perform destructive actions on their host or network. For more\ninformation on how OpenAI sandboxes its code, see the paper "Evaluating Large\nLanguage Models Trained on Code" (https://arxiv.org/abs/2107.03374).\n\nOnce you have read this disclaimer and taken appropriate precautions,\nset the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this\nwith:\n\n>>> import os\n>>> os.environ["HF_ALLOW_CODE_EVAL"] = "1"\n\n################################################################################\\n' UpperCAmelCase_ : List[str] = 'The MIT License\n\nCopyright (c) OpenAI (https://openai.com)\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the "Software"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in\nall copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\nTHE SOFTWARE.' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[Any]: return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/openai/human-eval' , codebase_urls=['https://github.com/openai/human-eval'] , reference_urls=['https://github.com/openai/human-eval'] , license=_LICENSE , ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple=[1, 1_0, 1_0_0] , SCREAMING_SNAKE_CASE__ : str=4 , SCREAMING_SNAKE_CASE__ : List[Any]=3.0 ) -> Union[str, Any]: if os.getenv('HF_ALLOW_CODE_EVAL' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('This metric is currently not supported on Windows.' ) with ThreadPoolExecutor(max_workers=SCREAMING_SNAKE_CASE__ ) as executor: a_ : str = [] a_ : Optional[int] = Counter() a_ : Union[str, Any] = 0 a_ : str = defaultdict(SCREAMING_SNAKE_CASE__ ) for task_id, (candidates, test_case) in enumerate(zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ): for candidate in candidates: a_ : Dict = candidate + '\n' + test_case a_ : str = (test_program, timeout, task_id, completion_id[task_id]) a_ : str = executor.submit(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ ) futures.append(SCREAMING_SNAKE_CASE__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(SCREAMING_SNAKE_CASE__ ): a_ : str = future.result() results[result["task_id"]].append((result['completion_id'], result) ) a_ , a_ : Union[str, Any] = [], [] for result in results.values(): result.sort() a_ : Optional[Any] = [r[1]['passed'] for r in result] total.append(len(SCREAMING_SNAKE_CASE__ ) ) correct.append(sum(SCREAMING_SNAKE_CASE__ ) ) a_ : Optional[int] = np.array(SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = np.array(SCREAMING_SNAKE_CASE__ ) a_ : Dict = k a_ : List[Any] = {F"""pass@{k}""": estimate_pass_at_k(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def SCREAMING_SNAKE_CASE_ ( __A : Optional[int] , __A : List[str] , __A : int ) -> Optional[int]: """simple docstring""" def estimator(__A : int , __A : int , __A : int ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(__A , __A ): a_ : Union[str, Any] = itertools.repeat(__A , len(__A ) ) else: assert len(__A ) == len(__A ) a_ : Union[str, Any] = iter(__A ) return np.array([estimator(int(__A ) , int(__A ) , __A ) for n, c in zip(__A , __A )] )
32
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __A : Tuple = { '''configuration_convbert''': ['''CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ConvBertConfig''', '''ConvBertOnnxConfig'''], '''tokenization_convbert''': ['''ConvBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = ['''ConvBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ '''CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ConvBertForMaskedLM''', '''ConvBertForMultipleChoice''', '''ConvBertForQuestionAnswering''', '''ConvBertForSequenceClassification''', '''ConvBertForTokenClassification''', '''ConvBertLayer''', '''ConvBertModel''', '''ConvBertPreTrainedModel''', '''load_tf_weights_in_convbert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = [ '''TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFConvBertForMaskedLM''', '''TFConvBertForMultipleChoice''', '''TFConvBertForQuestionAnswering''', '''TFConvBertForSequenceClassification''', '''TFConvBertForTokenClassification''', '''TFConvBertLayer''', '''TFConvBertModel''', '''TFConvBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __A : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
0
'''simple docstring''' import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class _a ( unittest.TestCase ): def __init__( self : int , lowercase : Union[str, Any] , lowercase : Dict=7 , lowercase : Union[str, Any]=3 , lowercase : int=30 , lowercase : List[str]=400 , lowercase : List[Any]=True , lowercase : Union[str, Any]=None , lowercase : Union[str, Any]=True , lowercase : Any=[0.5, 0.5, 0.5] , lowercase : int=[0.5, 0.5, 0.5] , lowercase : List[Any]=True , lowercase : Tuple=1 / 255 , lowercase : List[str]=True , ): '''simple docstring''' UpperCAmelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1_333} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_pad def A ( self : List[str] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Union[str, Any] , lowercase : Dict , lowercase : Optional[Any]=False ): '''simple docstring''' if not batched: UpperCAmelCase = image_inputs[0] if isinstance(lowercase , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] if w < h: UpperCAmelCase = int(self.size['''shortest_edge'''] * h / w ) UpperCAmelCase = self.size['''shortest_edge'''] elif w > h: UpperCAmelCase = self.size['''shortest_edge'''] UpperCAmelCase = int(self.size['''shortest_edge'''] * w / h ) else: UpperCAmelCase = self.size['''shortest_edge'''] UpperCAmelCase = self.size['''shortest_edge'''] else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(lowercase , key=lambda lowercase : item[0] )[0] UpperCAmelCase = max(lowercase , key=lambda lowercase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _a ( __a , unittest.TestCase ): __a : List[Any] = ConditionalDetrImageProcessor if is_vision_available() else None def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = ConditionalDetrImageProcessingTester(self ) @property def A ( self : int ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , '''image_mean''' ) ) self.assertTrue(hasattr(lowercase , '''image_std''' ) ) self.assertTrue(hasattr(lowercase , '''do_normalize''' ) ) self.assertTrue(hasattr(lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(lowercase , '''size''' ) ) def A ( self : Dict ): '''simple docstring''' UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1_333} ) self.assertEqual(image_processor.do_pad , lowercase ) UpperCAmelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowercase ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , lowercase ) def A ( self : Union[str, Any] ): '''simple docstring''' pass def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) UpperCAmelCase = image_processing(lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(lowercase , return_tensors='''pt''' ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(lowercase , return_tensors='''pt''' ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def A ( self : Dict ): '''simple docstring''' UpperCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: UpperCAmelCase = json.loads(f.read() ) UpperCAmelCase = {'''image_id''': 39_769, '''annotations''': target} # encode them UpperCAmelCase = ConditionalDetrImageProcessor.from_pretrained('''microsoft/conditional-detr-resnet-50''' ) UpperCAmelCase = image_processing(images=lowercase , annotations=lowercase , return_tensors='''pt''' ) # verify pixel values UpperCAmelCase = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , lowercase ) UpperCAmelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , lowercase , atol=1E-4 ) ) # verify area UpperCAmelCase = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , lowercase ) ) # verify boxes UpperCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , lowercase ) UpperCAmelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , lowercase , atol=1E-3 ) ) # verify image_id UpperCAmelCase = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , lowercase ) ) # verify is_crowd UpperCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , lowercase ) ) # verify class_labels UpperCAmelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , lowercase ) ) # verify orig_size UpperCAmelCase = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , lowercase ) ) # verify size UpperCAmelCase = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , lowercase ) ) @slow def A ( self : str ): '''simple docstring''' UpperCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: UpperCAmelCase = json.loads(f.read() ) UpperCAmelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 39_769, '''segments_info''': target} UpperCAmelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them UpperCAmelCase = ConditionalDetrImageProcessor(format='''coco_panoptic''' ) UpperCAmelCase = image_processing(images=lowercase , annotations=lowercase , masks_path=lowercase , return_tensors='''pt''' ) # verify pixel values UpperCAmelCase = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , lowercase ) UpperCAmelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , lowercase , atol=1E-4 ) ) # verify area UpperCAmelCase = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , lowercase ) ) # verify boxes UpperCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , lowercase ) UpperCAmelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , lowercase , atol=1E-3 ) ) # verify image_id UpperCAmelCase = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , lowercase ) ) # verify is_crowd UpperCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , lowercase ) ) # verify class_labels UpperCAmelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , lowercase ) ) # verify masks UpperCAmelCase = 822_873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , lowercase ) # verify orig_size UpperCAmelCase = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , lowercase ) ) # verify size UpperCAmelCase = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , lowercase ) )
34
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase__ ( lowercase ): def __init__( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : str = dataset _UpperCamelCase : Optional[Any] = process _UpperCamelCase : Optional[Any] = params def __len__( self : Tuple ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.dataset[i] _UpperCamelCase : Dict = self.process(lowerCamelCase__ ,**self.params ) return processed class lowercase__ ( lowercase ): def __init__( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int]=None ): '''simple docstring''' _UpperCamelCase : Optional[int] = loader _UpperCamelCase : Tuple = infer _UpperCamelCase : List[str] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _UpperCamelCase : Any = None _UpperCamelCase : Union[str, Any] = loader_batch_size # Internal bookkeeping _UpperCamelCase : Optional[Any] = None _UpperCamelCase : str = None def __len__( self : List[str] ): '''simple docstring''' return len(self.loader ) def __iter__( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = iter(self.loader ) return self def UpperCamelCase_ ( self : Any ): '''simple docstring''' if isinstance(self._loader_batch_data ,torch.Tensor ): # Batch data is simple tensor, just fetch the slice _UpperCamelCase : Union[str, Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _UpperCamelCase : Union[str, Any] = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Convert ModelOutput to tuple first _UpperCamelCase : str = element.to_tuple() if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Dict = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _UpperCamelCase : Optional[int] = None elif isinstance(element[self._loader_batch_index] ,torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] ,np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : Optional[Any] = np.expand_dims(element[self._loader_batch_index] ,0 ) else: # This is typically a list, so no need to `unsqueeze`. _UpperCamelCase : Union[str, Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _UpperCamelCase : Optional[int] = self._loader_batch_data.__class__(lowerCamelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _UpperCamelCase : Tuple = next(self.iterator ) _UpperCamelCase : List[str] = self.infer(lowerCamelCase__ ,**self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : List[Any] = processed else: _UpperCamelCase : List[Any] = list(processed.keys() )[0] _UpperCamelCase : Optional[int] = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : int = len(lowerCamelCase__ ) else: _UpperCamelCase : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : int = observed_batch_size # Setting internal index to unwrap the batch _UpperCamelCase : Dict = processed _UpperCamelCase : str = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase__ ( lowercase ): def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __iter__( self : Dict ): '''simple docstring''' _UpperCamelCase : str = iter(self.loader ) _UpperCamelCase : List[str] = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: _UpperCamelCase : Tuple = self.infer(next(self.iterator ) ,**self.params ) try: # Try to return next item _UpperCamelCase : Optional[Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) _UpperCamelCase : int = next(self.subiterator ) return processed class lowercase__ ( lowercase ): def __iter__( self : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : Dict = self.loader_batch_item() _UpperCamelCase : List[str] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator while not is_last: _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) if self.loader_batch_size is not None: if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : str = processed else: _UpperCamelCase : Any = list(processed.keys() )[0] _UpperCamelCase : Tuple = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Dict = len(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : Any = observed_batch_size _UpperCamelCase : List[Any] = processed _UpperCamelCase : int = 0 while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : List[Any] = self.loader_batch_item() _UpperCamelCase : Optional[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator else: _UpperCamelCase : Any = processed _UpperCamelCase : List[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) return accumulator class lowercase__ ( lowercase ): def __init__( self : Tuple ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : str = key def __len__( self : Dict ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : Optional[Any] = keya _UpperCamelCase : str = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : List[str] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
83
0
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[int]: try: with open(_lowerCAmelCase , """rb""" ) as flax_state_f: snake_case__ : Any = from_bytes(_lowerCAmelCase , flax_state_f.read() ) except UnpicklingError as e: try: with open(_lowerCAmelCase ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(_lowerCAmelCase , _lowerCAmelCase ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> int: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights snake_case__ : Optional[int] = flatten_dict(jax.tree_util.tree_map(lambda _lowerCAmelCase : x.dtype == jnp.bfloataa , _lowerCAmelCase ) ).values() if any(_lowerCAmelCase ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) snake_case__ : Optional[Any] = jax.tree_util.tree_map( lambda _lowerCAmelCase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , _lowerCAmelCase ) snake_case__ : Optional[int] = """""" snake_case__ : Any = flatten_dict(_lowerCAmelCase , sep=""".""" ) snake_case__ : Union[str, Any] = pt_model.state_dict() # keep track of unexpected & missing keys snake_case__ : Any = [] snake_case__ : List[Any] = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): snake_case__ : str = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: snake_case__ : Dict = flax_key_tuple_array[:-1] + ["""weight"""] snake_case__ : List[Any] = jnp.transpose(_lowerCAmelCase , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": snake_case__ : str = flax_key_tuple_array[:-1] + ["""weight"""] snake_case__ : Dict = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": snake_case__ : Dict = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(_lowerCAmelCase ): snake_case__ : int = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) snake_case__ : List[Any] = """.""".join(_lowerCAmelCase ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " f"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict snake_case__ : Tuple = np.asarray(_lowerCAmelCase ) if not isinstance(_lowerCAmelCase , np.ndarray ) else flax_tensor snake_case__ : Optional[int] = torch.from_numpy(_lowerCAmelCase ) # remove from missing keys missing_keys.remove(_lowerCAmelCase ) else: # weight is not expected by PyTorch model unexpected_keys.append(_lowerCAmelCase ) pt_model.load_state_dict(_lowerCAmelCase ) # re-transform missing_keys to list snake_case__ : Tuple = list(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" f" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" f" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" f" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(_lowerCAmelCase ) > 0: logger.warning( f"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" f" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" """ use it for predictions and inference.""" ) return pt_model
35
'''simple docstring''' import os from datetime import datetime as dt from github import Github snake_case_ : Any = [ 'good first issue', 'good second issue', 'good difficult issue', 'enhancement', 'new pipeline/model', 'new scheduler', 'wip', ] def A__ ( ): _UpperCamelCase : Tuple = Github(os.environ['GITHUB_TOKEN'] ) _UpperCamelCase : List[Any] = g.get_repo('huggingface/diffusers' ) _UpperCamelCase : List[Any] = repo.get_issues(state='open' ) for issue in open_issues: _UpperCamelCase : Dict = sorted(issue.get_comments() , key=lambda UpperCAmelCase_ : i.created_at , reverse=UpperCAmelCase_ ) _UpperCamelCase : List[str] = comments[0] if len(UpperCAmelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='closed' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='open' ) issue.remove_from_labels('stale' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) issue.add_to_labels('stale' ) if __name__ == "__main__": main()
83
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A ( ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ArgumentParser( description=( "PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCamelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCamelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCamelCase ) return parser.parse_args() def A ( ): '''simple docstring''' _lowerCAmelCase : List[str] = parse_args() # Import training_script as a module. _lowerCAmelCase : str = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) _lowerCAmelCase : List[Any] = script_fpath.stem _lowerCAmelCase : List[str] = importlib.import_module(_lowerCamelCase ) # Patch sys.argv _lowerCAmelCase : List[str] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
36
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase ) , """Tatoeba directory does not exist.""" ) class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Dict = self.resolver.write_model_card('opus-mt-he-en' ,dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
0
'''simple docstring''' import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = LongformerTokenizer __lowercase : Union[str, Any] = True __lowercase : Any = LongformerTokenizerFast __lowercase : int = True def UpperCAmelCase_ ( self ) -> List[str]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : str = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] lowerCAmelCase__ : Tuple = dict(zip(__UpperCAmelCase ,range(len(__UpperCAmelCase ) ) ) ) lowerCAmelCase__ : int = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] lowerCAmelCase__ : Union[str, Any] = {"""unk_token""": """<unk>"""} lowerCAmelCase__ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCAmelCase__ : int = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__UpperCAmelCase ) ) def UpperCAmelCase_ ( self ,**__UpperCAmelCase ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**__UpperCAmelCase ) def UpperCAmelCase_ ( self ,**__UpperCAmelCase ) -> List[str]: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Tuple: lowerCAmelCase__ : Union[str, Any] = """lower newer""" lowerCAmelCase__ : Union[str, Any] = """lower newer""" return input_text, output_text def UpperCAmelCase_ ( self ) -> Any: lowerCAmelCase__ : Optional[Any] = self.tokenizer_class(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) lowerCAmelCase__ : Optional[Any] = """lower newer""" lowerCAmelCase__ : List[Any] = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] lowerCAmelCase__ : Any = tokenizer.tokenize(__UpperCAmelCase ) # , add_prefix_space=True) self.assertListEqual(__UpperCAmelCase ,__UpperCAmelCase ) lowerCAmelCase__ : List[str] = tokens + [tokenizer.unk_token] lowerCAmelCase__ : int = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) ,__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : Optional[Any] = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""" ,add_special_tokens=__UpperCAmelCase ) ,[0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""" ,add_special_tokens=__UpperCAmelCase ) ,[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] ,) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: lowerCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained("""allenai/longformer-base-4096""" ) lowerCAmelCase__ : List[str] = tokenizer.encode("""sequence builders""" ,add_special_tokens=__UpperCAmelCase ) lowerCAmelCase__ : Any = tokenizer.encode("""multi-sequence build""" ,add_special_tokens=__UpperCAmelCase ) lowerCAmelCase__ : int = tokenizer.encode( """sequence builders""" ,add_special_tokens=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = tokenizer.encode( """sequence builders""" ,"""multi-sequence build""" ,add_special_tokens=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase ,__UpperCAmelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Union[str, Any] = self.get_tokenizer() lowerCAmelCase__ : List[str] = """Encode this sequence.""" lowerCAmelCase__ : Tuple = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments lowerCAmelCase__ : str = tokenizer.encode(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(__UpperCAmelCase ,__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = tokenizer.encode(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(__UpperCAmelCase ,__UpperCAmelCase ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) lowerCAmelCase__ : int = tokenizer.encode(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) lowerCAmelCase__ : List[str] = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(__UpperCAmelCase ,__UpperCAmelCase ) # Testing spaces after special tokens lowerCAmelCase__ : Optional[int] = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase )} ) # mask token has a left space lowerCAmelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = """Encode <mask> sequence""" lowerCAmelCase__ : int = """Encode <mask>sequence""" lowerCAmelCase__ : str = tokenizer.encode(__UpperCAmelCase ) lowerCAmelCase__ : int = encoded.index(__UpperCAmelCase ) lowerCAmelCase__ : Any = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(__UpperCAmelCase ,__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode(__UpperCAmelCase ) lowerCAmelCase__ : str = encoded.index(__UpperCAmelCase ) lowerCAmelCase__ : List[str] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(__UpperCAmelCase ,__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[str]: pass def UpperCAmelCase_ ( self ) -> Any: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Dict = self.tokenizer_class.from_pretrained(__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = """A, <mask> AllenNLP sentence.""" lowerCAmelCase__ : int = tokenizer_r.encode_plus(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ,return_token_type_ids=__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = tokenizer_p.encode_plus(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ,return_token_type_ids=__UpperCAmelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) ,sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) ,sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) ,) lowerCAmelCase__ : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) lowerCAmelCase__ : List[str] = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] ,[0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] ,[0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( __UpperCAmelCase ,["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( __UpperCAmelCase ,["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def UpperCAmelCase_ ( self ) -> int: for trim_offsets, add_prefix_space in itertools.product([True, False] ,repeat=2 ): lowerCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained( self.tmpdirname ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : int = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) lowerCAmelCase__ : List[str] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""] ,__UpperCAmelCase ) self.assertEqual(post_processor_state["""add_prefix_space"""] ,__UpperCAmelCase ) self.assertEqual(post_processor_state["""trim_offsets"""] ,__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase__ : Dict = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` lowerCAmelCase__ : Any = F"""{text_of_1_token} {text_of_1_token}""" lowerCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : List[str] = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(__UpperCAmelCase ) + 1, len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : Optional[int] = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : Any = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(__UpperCAmelCase ) + 1, len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(__UpperCAmelCase ), len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : Tuple = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(__UpperCAmelCase ), len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : int = F""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) lowerCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : Any = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(1, 1 + len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(__UpperCAmelCase ) + 1, 1 + len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : str = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, 1 + len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(__UpperCAmelCase ), 1 + len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,) lowerCAmelCase__ : Optional[Any] = self.rust_tokenizer_class.from_pretrained( __UpperCAmelCase ,use_fast=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,trim_offsets=__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = tokenizer_r(__UpperCAmelCase ,return_offsets_mapping=__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, 1 + len(__UpperCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(__UpperCAmelCase ), 1 + len(__UpperCAmelCase ) + 1 + len(__UpperCAmelCase )) ,)
37
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
0
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class _SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : Dict ): super().__init__() UpperCamelCase :List[str] = nn.Linear(3 , 4 ) UpperCamelCase :Dict = nn.BatchNormad(4 ) UpperCamelCase :Optional[Any] = nn.Linear(4 , 5 ) def _A ( self : Tuple , __lowerCamelCase : List[str] ): return self.lineara(self.batchnorm(self.lineara(__lowerCamelCase ) ) ) class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _A ( self : Tuple ): UpperCamelCase :List[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowerCamelCase , model.state_dict() ) UpperCamelCase :Dict = os.path.join(__lowerCamelCase , """index.json""" ) self.assertTrue(os.path.isfile(__lowerCamelCase ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: UpperCamelCase :Union[str, Any] = os.path.join(__lowerCamelCase , F"""{key}.dat""" ) self.assertTrue(os.path.isfile(__lowerCamelCase ) ) # TODO: add tests on the fact weights are properly loaded def _A ( self : Optional[Any] ): UpperCamelCase :List[str] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: UpperCamelCase :Optional[Any] = torch.randn(2 , 3 , dtype=__lowerCamelCase ) with TemporaryDirectory() as tmp_dir: UpperCamelCase :Dict = offload_weight(__lowerCamelCase , """weight""" , __lowerCamelCase , {} ) UpperCamelCase :Optional[Any] = os.path.join(__lowerCamelCase , """weight.dat""" ) self.assertTrue(os.path.isfile(__lowerCamelCase ) ) self.assertDictEqual(__lowerCamelCase , {"""weight""": {"""shape""": [2, 3], """dtype""": str(__lowerCamelCase ).split(""".""" )[1]}} ) UpperCamelCase :str = load_offloaded_weight(__lowerCamelCase , index["""weight"""] ) self.assertTrue(torch.equal(__lowerCamelCase , __lowerCamelCase ) ) def _A ( self : str ): UpperCamelCase :List[str] = ModelForTest() UpperCamelCase :Optional[int] = model.state_dict() UpperCamelCase :Optional[Any] = {k: v for k, v in state_dict.items() if """linear2""" not in k} UpperCamelCase :List[str] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :Optional[int] = OffloadedWeightsLoader(state_dict=__lowerCamelCase , save_folder=__lowerCamelCase ) # Every key is there with the right value self.assertEqual(sorted(__lowerCamelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowerCamelCase , weight_map[key] ) ) UpperCamelCase :Tuple = {k: v for k, v in state_dict.items() if """weight""" in k} UpperCamelCase :List[Any] = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :Optional[int] = OffloadedWeightsLoader(state_dict=__lowerCamelCase , save_folder=__lowerCamelCase ) # Every key is there with the right value self.assertEqual(sorted(__lowerCamelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowerCamelCase , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowerCamelCase , __lowerCamelCase ) # Duplicates are removed UpperCamelCase :Optional[int] = OffloadedWeightsLoader(state_dict=__lowerCamelCase , save_folder=__lowerCamelCase ) # Every key is there with the right value self.assertEqual(sorted(__lowerCamelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowerCamelCase , weight_map[key] ) ) def _A ( self : int ): UpperCamelCase :str = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} UpperCamelCase :Optional[int] = extract_submodules_state_dict(__lowerCamelCase , ["""a.1""", """a.2"""] ) self.assertDictEqual(__lowerCamelCase , {"""a.1""": 0, """a.2""": 2} ) UpperCamelCase :List[Any] = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} UpperCamelCase :List[Any] = extract_submodules_state_dict(__lowerCamelCase , ["""a.1""", """a.2"""] ) self.assertDictEqual(__lowerCamelCase , {"""a.1.a""": 0, """a.2.a""": 2} )
38
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : Dict = 3 _UpperCamelCase : Any = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 1_5 == 0: result -= a a += 1 return result if __name__ == "__main__": print(F"""{solution() = }""")
83
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = tempfile.mkdtemp() _UpperCAmelCase = BlipImageProcessor() _UpperCAmelCase = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-BertModel' ) _UpperCAmelCase = BlipProcessor(UpperCAmelCase , UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase ).tokenizer def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase ).image_processor def UpperCamelCase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _UpperCAmelCase = [Image.fromarray(np.moveaxis(UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = BlipProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _UpperCAmelCase = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _UpperCAmelCase = self.get_image_processor(do_normalize=UpperCAmelCase , padding_value=1.0 ) _UpperCAmelCase = BlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_image_processor() _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) _UpperCAmelCase = self.prepare_image_inputs() _UpperCAmelCase = image_processor(UpperCAmelCase , return_tensors='np' ) _UpperCAmelCase = processor(images=UpperCAmelCase , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_image_processor() _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) _UpperCAmelCase = 'lower newer' _UpperCAmelCase = processor(text=UpperCAmelCase ) _UpperCAmelCase = tokenizer(UpperCAmelCase , return_token_type_ids=UpperCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_image_processor() _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) _UpperCAmelCase = 'lower newer' _UpperCAmelCase = self.prepare_image_inputs() _UpperCAmelCase = processor(text=UpperCAmelCase , images=UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'input_ids', 'attention_mask'] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase ): processor() def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_image_processor() _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) _UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _UpperCAmelCase = processor.batch_decode(UpperCAmelCase ) _UpperCAmelCase = tokenizer.batch_decode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_image_processor() _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) _UpperCAmelCase = 'lower newer' _UpperCAmelCase = self.prepare_image_inputs() _UpperCAmelCase = processor(text=UpperCAmelCase , images=UpperCAmelCase ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'input_ids', 'attention_mask'] )
39
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
83
0
"""simple docstring""" from __future__ import annotations class _A : """simple docstring""" def __init__( self : List[str] , __UpperCAmelCase : int = 0): a : Tuple = key def __snake_case ( self : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : int): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) a : Dict = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(__UpperCAmelCase) ^ key) for ch in content] def __snake_case ( self : int , __UpperCAmelCase : str , __UpperCAmelCase : int): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) a : Optional[Any] = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(__UpperCAmelCase) ^ key) for ch in content] def __snake_case ( self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : int = 0): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) a : List[Any] = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned a : Any = "" for ch in content: ans += chr(ord(__UpperCAmelCase) ^ key) return ans def __snake_case ( self : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : int = 0): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) a : Dict = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned a : str = "" for ch in content: ans += chr(ord(__UpperCAmelCase) ^ key) return ans def __snake_case ( self : int , __UpperCAmelCase : str , __UpperCAmelCase : int = 0): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) try: with open(__UpperCAmelCase) as fin, open("encrypt.out" , "w+") as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(__UpperCAmelCase , __UpperCAmelCase)) except OSError: return False return True def __snake_case ( self : Any , __UpperCAmelCase : str , __UpperCAmelCase : int): assert isinstance(__UpperCAmelCase , __UpperCAmelCase) and isinstance(__UpperCAmelCase , __UpperCAmelCase) try: with open(__UpperCAmelCase) as fin, open("decrypt.out" , "w+") as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(__UpperCAmelCase , __UpperCAmelCase)) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
40
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') snake_case_ : Any = logging.getLogger(__name__) @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) lowercase__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class lowercase__ : lowercase__ = field(default=lowercase , metadata={"""help""": """The input training data file (a text file)."""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowercase__ = field( default=lowercase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def UpperCamelCase_ ( self : str ): '''simple docstring''' if self.train_file is not None: _UpperCamelCase : List[Any] = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _UpperCamelCase : Union[str, Any] = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowercase__ : lowercase__ = 42 lowercase__ = True lowercase__ = None lowercase__ = None def __call__( self : Optional[Any] ,lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = 'label' if 'label' in features[0].keys() else 'labels' _UpperCamelCase : List[Any] = [feature.pop(lowerCamelCase__ ) for feature in features] _UpperCamelCase : Dict = len(lowerCamelCase__ ) _UpperCamelCase : List[str] = len(features[0]['input_ids'] ) _UpperCamelCase : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(lowerCamelCase__ )] for feature in features ] _UpperCamelCase : str = list(chain(*lowerCamelCase__ ) ) _UpperCamelCase : Tuple = self.tokenizer.pad( lowerCamelCase__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='pt' ,) # Un-flatten _UpperCamelCase : str = {k: v.view(lowerCamelCase__ ,lowerCamelCase__ ,-1 ) for k, v in batch.items()} # Add back labels _UpperCamelCase : Optional[int] = torch.tensor(lowerCamelCase__ ,dtype=torch.intaa ) return batch def A__ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCamelCase : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Optional[int] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , UpperCAmelCase_ , UpperCAmelCase_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(UpperCAmelCase_ ) datasets.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.set_verbosity(UpperCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCamelCase : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCamelCase : List[str] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _UpperCamelCase : Optional[int] = {} if data_args.train_file is not None: _UpperCamelCase : Tuple = data_args.train_file if data_args.validation_file is not None: _UpperCamelCase : Tuple = data_args.validation_file _UpperCamelCase : Any = data_args.train_file.split('.' )[-1] _UpperCamelCase : Union[str, Any] = load_dataset( UpperCAmelCase_ , data_files=UpperCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _UpperCamelCase : List[str] = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=UpperCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _UpperCamelCase : Any = [f'ending{i}' for i in range(4 )] _UpperCamelCase : int = 'sent1' _UpperCamelCase : List[str] = 'sent2' if data_args.max_seq_length is None: _UpperCamelCase : int = tokenizer.model_max_length if max_seq_length > 1_0_2_4: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) _UpperCamelCase : int = 1_0_2_4 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) _UpperCamelCase : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(UpperCAmelCase_ ): _UpperCamelCase : str = [[context] * 4 for context in examples[context_name]] _UpperCamelCase : Optional[Any] = examples[question_header_name] _UpperCamelCase : Tuple = [ [f'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(UpperCAmelCase_ ) ] # Flatten out _UpperCamelCase : Optional[int] = list(chain(*UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = list(chain(*UpperCAmelCase_ ) ) # Tokenize _UpperCamelCase : Tuple = tokenizer( UpperCAmelCase_ , UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(UpperCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _UpperCamelCase : Optional[Any] = raw_datasets['train'] if data_args.max_train_samples is not None: _UpperCamelCase : Tuple = min(len(UpperCAmelCase_ ) , data_args.max_train_samples ) _UpperCamelCase : Tuple = train_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): _UpperCamelCase : Union[str, Any] = train_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _UpperCamelCase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _UpperCamelCase : Union[str, Any] = min(len(UpperCAmelCase_ ) , data_args.max_eval_samples ) _UpperCamelCase : str = eval_dataset.select(range(UpperCAmelCase_ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): _UpperCamelCase : Dict = eval_dataset.map( UpperCAmelCase_ , batched=UpperCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _UpperCamelCase : List[Any] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=UpperCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Union[str, Any] = eval_predictions _UpperCamelCase : List[str] = np.argmax(UpperCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer _UpperCamelCase : Optional[int] = Trainer( model=UpperCAmelCase_ , args=UpperCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=UpperCAmelCase_ , data_collator=UpperCAmelCase_ , compute_metrics=UpperCAmelCase_ , ) # Training if training_args.do_train: _UpperCamelCase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _UpperCamelCase : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCamelCase : int = last_checkpoint _UpperCamelCase : List[str] = trainer.train(resume_from_checkpoint=UpperCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload _UpperCamelCase : Union[str, Any] = train_result.metrics _UpperCamelCase : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCAmelCase_ ) ) _UpperCamelCase : Optional[Any] = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('train' , UpperCAmelCase_ ) trainer.save_metrics('train' , UpperCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _UpperCamelCase : List[Any] = trainer.evaluate() _UpperCamelCase : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(UpperCAmelCase_ ) _UpperCamelCase : int = min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) trainer.log_metrics('eval' , UpperCAmelCase_ ) trainer.save_metrics('eval' , UpperCAmelCase_ ) _UpperCamelCase : Optional[int] = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**UpperCAmelCase_ ) else: trainer.create_model_card(**UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
83
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class _lowercase : def __init__( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Optional[Any]=13 , UpperCamelCase__: Dict=7 , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=True , UpperCamelCase__: List[str]=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: List[str]=99 , UpperCamelCase__: Dict=32 , UpperCamelCase__: int=2 , UpperCamelCase__: Any=4 , UpperCamelCase__: Optional[int]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[str]=0.1 , UpperCamelCase__: Dict=512 , UpperCamelCase__: Union[str, Any]=16 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: int=0.02 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Optional[Any]=4 , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: List[Any]=0 , ): lowerCamelCase__ : str = parent lowerCamelCase__ : Tuple = batch_size lowerCamelCase__ : Optional[int] = seq_length lowerCamelCase__ : Dict = is_training lowerCamelCase__ : Optional[int] = use_input_mask lowerCamelCase__ : Optional[Any] = use_token_type_ids lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Tuple = vocab_size lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : List[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[int] = intermediate_size lowerCamelCase__ : List[str] = hidden_act lowerCamelCase__ : Dict = hidden_dropout_prob lowerCamelCase__ : Dict = attention_probs_dropout_prob lowerCamelCase__ : str = max_position_embeddings lowerCamelCase__ : List[str] = type_vocab_size lowerCamelCase__ : Optional[Any] = type_sequence_label_size lowerCamelCase__ : Optional[Any] = initializer_range lowerCamelCase__ : int = num_labels lowerCamelCase__ : List[Any] = num_choices lowerCamelCase__ : int = scope lowerCamelCase__ : Tuple = projection_dim def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : Dict = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py lowerCamelCase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase__ : Tuple = None if self.use_token_type_ids: lowerCamelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : int = None lowerCamelCase__ : Tuple = None if self.use_labels: lowerCamelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase__ : Union[str, Any] = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) lowerCamelCase__ : Tuple = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: List[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: str , UpperCamelCase__: int , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[Any] = TFDPRContextEncoder(config=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : str = model(UpperCamelCase__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[str] , UpperCamelCase__: List[str] , UpperCamelCase__: List[Any] , UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: Any ): lowerCamelCase__ : int = TFDPRQuestionEncoder(config=UpperCamelCase__ ) lowerCamelCase__ : Any = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: str , UpperCamelCase__: List[str] , UpperCamelCase__: Any , UpperCamelCase__: int , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): lowerCamelCase__ : str = TFDPRReader(config=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : Optional[Any] = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : str = config_and_inputs lowerCamelCase__ : Optional[int] = {"""input_ids""": input_ids} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) a = {"""feature-extraction""": TFDPRQuestionEncoder} if is_tf_available() else {} a = False a = False a = False a = False a = False def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = TFDPRModelTester(self ) lowerCamelCase__ : int = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Dict ): self.config_tester.run_common_tests() def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: Optional[int] ): for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : str = TFDPRContextEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Optional[int] = TFDPRContextEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Tuple = TFDPRQuestionEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : int = TFDPRReader.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) @require_tf class _lowercase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : List[str] = TFDPRQuestionEncoder.from_pretrained("""facebook/dpr-question_encoder-single-nq-base""" ) lowerCamelCase__ : Optional[Any] = tf.constant( [[101, 7_592, 1_010, 2_003, 2_026, 3_899, 10_140, 1_029, 102]] ) # [CLS] hello, is my dog cute? [SEP] lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ )[0] # embedding shape = (1, 768) # compare the actual values for a slice. lowerCamelCase__ : Union[str, Any] = tf.constant( [ [ 0.03_236_253, 0.12_753_335, 0.16_818_509, 0.00_279_786, 0.3_896_933, 0.24_264_945, 0.2_178_971, -0.02_335_227, -0.08_481_959, -0.14_324_117, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1e-4 ) )
41
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) lowercase__ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) lowercase__ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def A__ ( ): _UpperCamelCase : Optional[Any] = HfArgumentParser((ModelArguments,) ) ((_UpperCamelCase) , ) : Optional[int] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _UpperCamelCase : Any = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _UpperCamelCase : List[Any] = True _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : str = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=UpperCAmelCase_ , decoder_config=UpperCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _UpperCamelCase : str = decoder_config.decoder_start_token_id _UpperCamelCase : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _UpperCamelCase : int = decoder_config.bos_token_id if pad_token_id is None: _UpperCamelCase : Dict = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _UpperCamelCase : List[Any] = decoder_config.eos_token_id _UpperCamelCase : Dict = decoder_start_token_id _UpperCamelCase : int = pad_token_id _UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
83
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase : Any = logging.get_logger(__name__) lowercase : str = { "google/mobilenet_v1_1.0_224": "https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json", "google/mobilenet_v1_0.75_192": "https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """mobilenet_v1""" def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=2_24 , lowerCAmelCase_=1.0 , lowerCAmelCase_=8 , lowerCAmelCase_="relu6" , lowerCAmelCase_=True , lowerCAmelCase_=0.999 , lowerCAmelCase_=0.02 , lowerCAmelCase_=0.001 , **lowerCAmelCase_ , ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) if depth_multiplier <= 0: raise ValueError('depth_multiplier must be greater than zero.' ) _snake_case = num_channels _snake_case = image_size _snake_case = depth_multiplier _snake_case = min_depth _snake_case = hidden_act _snake_case = tf_padding _snake_case = classifier_dropout_prob _snake_case = initializer_range _snake_case = layer_norm_eps class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): """simple docstring""" return OrderedDict([('pixel_values', {0: 'batch'})] ) @property def lowerCamelCase ( self ): """simple docstring""" if self.task == "image-classification": return OrderedDict([('logits', {0: 'batch'})] ) else: return OrderedDict([('last_hidden_state', {0: 'batch'}), ('pooler_output', {0: 'batch'})] ) @property def lowerCamelCase ( self ): """simple docstring""" return 1E-4
42
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy snake_case_ : Dict = logging.get_logger(__name__) class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : float ,**lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : List[Any] = feature_size _UpperCamelCase : Any = sampling_rate _UpperCamelCase : Optional[Any] = padding_value _UpperCamelCase : Union[str, Any] = kwargs.pop('padding_side' ,'right' ) _UpperCamelCase : Dict = kwargs.pop('return_attention_mask' ,lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] ,lowerCamelCase__ : Union[bool, str, PaddingStrategy] = True ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : bool = False ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,lowerCamelCase__ : Optional[Union[str, TensorType]] = None ,): '''simple docstring''' # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowerCamelCase__ ,(list, tuple) ) and isinstance(processed_features[0] ,(dict, BatchFeature) ): _UpperCamelCase : int = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' F' to this method that includes {self.model_input_names[0]}, but you provided' F' {list(processed_features.keys() )}' ) _UpperCamelCase : List[Any] = processed_features[self.model_input_names[0]] _UpperCamelCase : Dict = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowerCamelCase__ ) == 0: if return_attention_mask: _UpperCamelCase : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch _UpperCamelCase : List[str] = required_input[0] if isinstance(lowerCamelCase__ ,(list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. _UpperCamelCase : List[str] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowerCamelCase__ ): _UpperCamelCase : Dict = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowerCamelCase__ ): _UpperCamelCase : Any = 'tf' elif is_torch_tensor(lowerCamelCase__ ): _UpperCamelCase : Optional[int] = 'pt' elif isinstance(lowerCamelCase__ ,(int, float, list, tuple, np.ndarray) ): _UpperCamelCase : int = 'np' else: raise ValueError( F'type of {first_element} unknown: {type(lowerCamelCase__ )}. ' 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] ,(int, float) ): _UpperCamelCase : Any = to_numpy(lowerCamelCase__ ) else: _UpperCamelCase : Any = [to_numpy(lowerCamelCase__ ) for v in value] # Convert padding_strategy in PaddingStrategy _UpperCamelCase : Optional[int] = self._get_padding_strategies(padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ) _UpperCamelCase : str = processed_features[self.model_input_names[0]] _UpperCamelCase : List[str] = len(lowerCamelCase__ ) if not all(len(lowerCamelCase__ ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) _UpperCamelCase : List[str] = [] for i in range(lowerCamelCase__ ): _UpperCamelCase : List[str] = {k: v[i] for k, v in processed_features.items()} # truncation _UpperCamelCase : List[str] = self._truncate( lowerCamelCase__ ,max_length=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,truncation=lowerCamelCase__ ,) truncated_inputs.append(lowerCamelCase__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length _UpperCamelCase : Union[str, Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) _UpperCamelCase : Any = PaddingStrategy.MAX_LENGTH _UpperCamelCase : Optional[Any] = {} for i in range(lowerCamelCase__ ): # padding _UpperCamelCase : Any = self._pad( truncated_inputs[i] ,max_length=lowerCamelCase__ ,padding_strategy=lowerCamelCase__ ,pad_to_multiple_of=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,) for key, value in outputs.items(): if key not in batch_outputs: _UpperCamelCase : Dict = [] if value.dtype is np.dtype(np.floataa ): _UpperCamelCase : Any = value.astype(np.floataa ) batch_outputs[key].append(lowerCamelCase__ ) return BatchFeature(lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' _UpperCamelCase : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: _UpperCamelCase : Optional[Any] = len(lowerCamelCase__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : str = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : str = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowerCamelCase__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: _UpperCamelCase : Tuple = np.ones(len(lowerCamelCase__ ) ,dtype=np.intaa ) if needs_to_be_padded: _UpperCamelCase : Dict = max_length - len(lowerCamelCase__ ) if self.padding_side == "right": if return_attention_mask: _UpperCamelCase : Optional[int] = np.pad( processed_features['attention_mask'] ,(0, difference) ) _UpperCamelCase : Union[str, Any] = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) _UpperCamelCase : List[Any] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: _UpperCamelCase : List[Any] = np.pad( processed_features['attention_mask'] ,(difference, 0) ) _UpperCamelCase : List[Any] = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) _UpperCamelCase : List[str] = np.pad( lowerCamelCase__ ,lowerCamelCase__ ,'constant' ,constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[Dict[str, np.ndarray], BatchFeature] ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[int] = None ,lowerCamelCase__ : Optional[bool] = None ,): '''simple docstring''' if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) _UpperCamelCase : int = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): _UpperCamelCase : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of _UpperCamelCase : Optional[int] = len(lowerCamelCase__ ) > max_length if needs_to_be_truncated: _UpperCamelCase : Dict = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: _UpperCamelCase : Optional[Any] = processed_features['attention_mask'][:max_length] return processed_features def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : int=False ,lowerCamelCase__ : Optional[Any]=None ): '''simple docstring''' # Get padding strategy if padding is not False: if padding is True: _UpperCamelCase : Optional[Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Tuple = PaddingStrategy(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = padding else: _UpperCamelCase : List[Any] = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
83
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :int = tempfile.mkdtemp() # fmt: off __UpperCamelCase :int = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on __UpperCamelCase :Tuple = dict(zip(__lowercase , range(len(__lowercase)))) __UpperCamelCase :List[Any] = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] __UpperCamelCase :Optional[Any] = {'''unk_token''': '''<unk>'''} __UpperCamelCase :List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) __UpperCamelCase :List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: fp.write(json.dumps(__lowercase) + '''\n''') with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(__lowercase)) __UpperCamelCase :List[str] = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], '''image_std''': [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], } __UpperCamelCase :Any = os.path.join(self.tmpdirname , __lowercase) with open(self.image_processor_file , '''w''' , encoding='''utf-8''') as fp: json.dump(__lowercase , __lowercase) def UpperCamelCase__ ( self , **__lowercase) -> Any: return CLIPTokenizer.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self , **__lowercase) -> Any: return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self , **__lowercase) -> List[str]: return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self) -> List[str]: shutil.rmtree(self.tmpdirname) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Dict = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta)] __UpperCamelCase :int = [Image.fromarray(np.moveaxis(__lowercase , 0 , -1)) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Dict = self.get_tokenizer() __UpperCamelCase :str = self.get_rust_tokenizer() __UpperCamelCase :Optional[Any] = self.get_image_processor() __UpperCamelCase :List[Any] = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) processor_slow.save_pretrained(self.tmpdirname) __UpperCamelCase :Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=__lowercase) __UpperCamelCase :Union[str, Any] = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) processor_fast.save_pretrained(self.tmpdirname) __UpperCamelCase :Dict = CLIPSegProcessor.from_pretrained(self.tmpdirname) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab()) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab()) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab()) self.assertIsInstance(processor_slow.tokenizer , __lowercase) self.assertIsInstance(processor_fast.tokenizer , __lowercase) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string()) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor_slow.image_processor , __lowercase) self.assertIsInstance(processor_fast.image_processor , __lowercase) def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) __UpperCamelCase :int = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''') __UpperCamelCase :List[Any] = self.get_image_processor(do_normalize=__lowercase , padding_value=1.0) __UpperCamelCase :Optional[Any] = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__lowercase , padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer , __lowercase) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __lowercase) def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :Optional[int] = self.get_image_processor() __UpperCamelCase :Optional[Any] = self.get_tokenizer() __UpperCamelCase :Dict = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Dict = self.prepare_image_inputs() __UpperCamelCase :Union[str, Any] = image_processor(__lowercase , return_tensors='''np''') __UpperCamelCase :str = processor(images=__lowercase , return_tensors='''np''') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = self.get_image_processor() __UpperCamelCase :Dict = self.get_tokenizer() __UpperCamelCase :str = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Tuple = '''lower newer''' __UpperCamelCase :str = processor(text=__lowercase) __UpperCamelCase :Optional[Any] = tokenizer(__lowercase) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :str = self.get_image_processor() __UpperCamelCase :Tuple = self.get_tokenizer() __UpperCamelCase :Optional[Any] = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Optional[int] = '''lower newer''' __UpperCamelCase :str = self.prepare_image_inputs() __UpperCamelCase :Any = processor(text=__lowercase , images=__lowercase) self.assertListEqual(list(inputs.keys()) , ['''input_ids''', '''attention_mask''', '''pixel_values''']) # test if it raises when no input is passed with pytest.raises(__lowercase): processor() def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :str = self.get_image_processor() __UpperCamelCase :Tuple = self.get_tokenizer() __UpperCamelCase :List[Any] = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Optional[Any] = self.prepare_image_inputs() __UpperCamelCase :Optional[Any] = self.prepare_image_inputs() __UpperCamelCase :List[Any] = processor(images=__lowercase , visual_prompt=__lowercase) self.assertListEqual(list(inputs.keys()) , ['''pixel_values''', '''conditional_pixel_values''']) # test if it raises when no input is passed with pytest.raises(__lowercase): processor() def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Optional[Any] = self.get_image_processor() __UpperCamelCase :List[str] = self.get_tokenizer() __UpperCamelCase :Optional[Any] = CLIPSegProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :int = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __UpperCamelCase :Optional[Any] = processor.batch_decode(__lowercase) __UpperCamelCase :Any = tokenizer.batch_decode(__lowercase) self.assertListEqual(__lowercase , __lowercase)
43
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
0
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class __A : def __init__( self , a__ , a__=13 , a__=7 , a__=True , a__=True , a__=False , a__=True , a__=99 , a__=32 , a__=5 , a__=4 , a__=37 , a__="gelu" , a__=0.1 , a__=0.1 , a__=512 , a__=16 , a__=2 , a__=0.0_2 , a__=3 , a__=4 , a__=None , ): _lowerCAmelCase : List[str] = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : str = seq_length _lowerCAmelCase : List[Any] = is_training _lowerCAmelCase : Union[str, Any] = use_input_mask _lowerCAmelCase : List[str] = use_token_type_ids _lowerCAmelCase : Tuple = use_labels _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Union[str, Any] = hidden_size _lowerCAmelCase : Union[str, Any] = num_hidden_layers _lowerCAmelCase : Any = num_attention_heads _lowerCAmelCase : str = intermediate_size _lowerCAmelCase : Dict = hidden_act _lowerCAmelCase : List[Any] = hidden_dropout_prob _lowerCAmelCase : int = attention_probs_dropout_prob _lowerCAmelCase : Dict = max_position_embeddings _lowerCAmelCase : List[Any] = type_vocab_size _lowerCAmelCase : int = type_sequence_label_size _lowerCAmelCase : Union[str, Any] = initializer_range _lowerCAmelCase : List[str] = num_labels _lowerCAmelCase : Optional[int] = num_choices _lowerCAmelCase : Optional[Any] = scope def __A ( self ): _lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : Optional[int] = None if self.use_input_mask: _lowerCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase : List[str] = None if self.use_token_type_ids: _lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : int = None _lowerCAmelCase : Any = None _lowerCAmelCase : Optional[Any] = None if self.use_labels: _lowerCAmelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCAmelCase : Dict = ids_tensor([self.batch_size] , self.num_choices ) _lowerCAmelCase : List[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __A ( self ): return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a__ , initializer_range=self.initializer_range , ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ ): _lowerCAmelCase : Any = LlamaModel(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Dict = model(a__ , attention_mask=a__ ) _lowerCAmelCase : str = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Tuple = True _lowerCAmelCase : Tuple = LlamaModel(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : List[str] = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , ) _lowerCAmelCase : Optional[Any] = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , ) _lowerCAmelCase : Any = model(a__ , attention_mask=a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Tuple = LlamaForCausalLM(config=a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : int = model(a__ , attention_mask=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Optional[int] = True _lowerCAmelCase : Optional[int] = True _lowerCAmelCase : Tuple = LlamaForCausalLM(config=a__ ) model.to(a__ ) model.eval() # first forward pass _lowerCAmelCase : Union[str, Any] = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , use_cache=a__ , ) _lowerCAmelCase : Dict = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _lowerCAmelCase : Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCAmelCase : Any = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _lowerCAmelCase : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCAmelCase : Optional[int] = torch.cat([input_mask, next_mask] , dim=-1 ) _lowerCAmelCase : Tuple = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , output_hidden_states=a__ , )["""hidden_states"""][0] _lowerCAmelCase : Optional[Any] = model( a__ , attention_mask=a__ , encoder_hidden_states=a__ , encoder_attention_mask=a__ , past_key_values=a__ , output_hidden_states=a__ , )["""hidden_states"""][0] # select random slice _lowerCAmelCase : Optional[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCAmelCase : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCAmelCase : Optional[int] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a__ , a__ , atol=1e-3 ) ) def __A ( self ): _lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCAmelCase : int = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __A ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Optional[Any] = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () _UpperCamelCase : Dict = (LlamaForCausalLM,) if is_torch_available() else () _UpperCamelCase : Union[str, Any] = ( { "feature-extraction": LlamaModel, "text-classification": LlamaForSequenceClassification, "text-generation": LlamaForCausalLM, "zero-shot": LlamaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase : Optional[Any] = False _UpperCamelCase : Dict = False def __A ( self ): _lowerCAmelCase : Dict = LlamaModelTester(self ) _lowerCAmelCase : List[str] = ConfigTester(self , config_class=a__ , hidden_size=37 ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): _lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def __A ( self ): _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCAmelCase : List[str] = type self.model_tester.create_and_check_model(*a__ ) def __A ( self ): _lowerCAmelCase , _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : Optional[int] = 3 _lowerCAmelCase : Union[str, Any] = input_dict["""input_ids"""] _lowerCAmelCase : str = input_ids.ne(1 ).to(a__ ) _lowerCAmelCase : int = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCAmelCase : List[Any] = LlamaForSequenceClassification(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : int = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __A ( self ): _lowerCAmelCase , _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : int = 3 _lowerCAmelCase : str = """single_label_classification""" _lowerCAmelCase : Optional[int] = input_dict["""input_ids"""] _lowerCAmelCase : Union[str, Any] = input_ids.ne(1 ).to(a__ ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCAmelCase : Union[str, Any] = LlamaForSequenceClassification(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : Optional[Any] = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __A ( self ): _lowerCAmelCase , _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : List[str] = 3 _lowerCAmelCase : int = """multi_label_classification""" _lowerCAmelCase : Tuple = input_dict["""input_ids"""] _lowerCAmelCase : Dict = input_ids.ne(1 ).to(a__ ) _lowerCAmelCase : int = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _lowerCAmelCase : Union[str, Any] = LlamaForSequenceClassification(a__ ) model.to(a__ ) model.eval() _lowerCAmelCase : int = model(a__ , attention_mask=a__ , labels=a__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("""LLaMA buffers include complex numbers, which breaks this test""" ) def __A ( self ): pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def __A ( self , a__ ): _lowerCAmelCase , _lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : int = ids_tensor([1, 10] , config.vocab_size ) _lowerCAmelCase : int = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCAmelCase : List[str] = LlamaModel(a__ ) original_model.to(a__ ) original_model.eval() _lowerCAmelCase : List[str] = original_model(a__ ).last_hidden_state _lowerCAmelCase : Optional[Any] = original_model(a__ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCAmelCase : Tuple = {"""type""": scaling_type, """factor""": 1_0.0} _lowerCAmelCase : Optional[Any] = LlamaModel(a__ ) scaled_model.to(a__ ) scaled_model.eval() _lowerCAmelCase : Any = scaled_model(a__ ).last_hidden_state _lowerCAmelCase : Optional[Any] = scaled_model(a__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(a__ , a__ , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(a__ , a__ , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(a__ , a__ , atol=1e-5 ) ) @require_torch class __A ( unittest.TestCase ): @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __A ( self ): _lowerCAmelCase : Tuple = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : List[Any] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-7b-hf""" , device_map="""auto""" ) _lowerCAmelCase : Tuple = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 _lowerCAmelCase : Optional[int] = torch.tensor([[-6.6_5_5_0, -4.1_2_2_7, -4.9_8_5_9, -3.2_4_0_6, 0.8_2_6_2, -3.0_0_3_3, 1.2_9_6_4, -3.3_6_9_9]] ) torch.testing.assert_close(out.mean(-1 ) , a__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : List[str] = torch.tensor([-1_2.8_2_8_1, -7.4_4_5_3, -0.4_6_3_9, -8.0_6_2_5, -7.2_5_0_0, -8.0_0_0_0, -6.4_8_8_3, -7.7_6_9_5, -7.8_4_3_8, -7.0_3_1_2, -6.2_1_8_8, -7.1_3_2_8, -1.8_4_9_6, 1.9_9_6_1, -8.6_2_5_0, -6.7_2_2_7, -1_2.8_2_8_1, -6.9_4_9_2, -7.0_7_4_2, -7.7_8_5_2, -7.5_8_2_0, -7.9_0_6_2, -6.9_3_7_5, -7.9_8_0_5, -8.3_4_3_8, -8.1_5_6_2, -8.0_4_6_9, -7.6_2_5_0, -7.7_4_2_2, -7.3_3_9_8,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , a__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __A ( self ): _lowerCAmelCase : Optional[int] = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : List[Any] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-13b-hf""" , device_map="""auto""" ) _lowerCAmelCase : Tuple = model(torch.tensor(a__ ) ) # Expected mean on dim = -1 _lowerCAmelCase : Union[str, Any] = torch.tensor([[-2.0_6_2_2, -1.2_7_9_4, -1.1_6_3_8, -0.9_7_8_8, -1.4_6_0_3, -1.0_2_3_8, -1.7_8_9_3, -1.4_4_1_1]] ) torch.testing.assert_close(out.mean(-1 ) , a__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : List[str] = torch.tensor([-8.1_4_0_6, -8.0_5_4_7, 2.7_4_6_1, -1.2_3_4_4, -0.1_4_4_8, -1.8_2_6_2, -1.0_0_2_0, -1.8_1_5_4, -1.6_8_9_5, -1.8_5_1_6, -2.3_5_7_4, -0.9_2_7_7, 3.7_5_9_8, 6.5_7_4_2, -1.2_9_9_8, -0.1_1_7_7, -8.1_4_0_6, -2.9_6_8_8, -2.9_1_9_9, -3.1_6_9_9, -3.5_2_5_4, -2.3_5_5_5, -2.7_9_8_8, -3.4_1_4_1, -2.8_2_6_2, -4.5_1_9_5, -3.3_3_7_9, -3.3_1_6_4, -2.7_8_3_2, -3.0_2_7_3] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , a__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __A ( self ): _lowerCAmelCase : int = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : Dict = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-13b-chat-hf""" , device_map="""auto""" ) _lowerCAmelCase : Tuple = model(torch.tensor(a__ ) ) # Expected mean on dim = -1 _lowerCAmelCase : Optional[Any] = torch.tensor([[-0.8_5_6_2, -1.8_5_2_0, -0.7_5_5_1, -0.4_1_6_2, -1.5_1_6_1, -1.2_0_3_8, -2.4_8_2_3, -2.3_2_5_4]] ) torch.testing.assert_close(out.mean(-1 ) , a__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : List[str] = torch.tensor([-2.2_2_2_7, 4.8_8_2_8, 0.9_0_2_3, -0.4_5_7_8, -0.7_8_7_1, -0.1_0_3_3, -0.6_2_2_1, -0.5_7_8_6, -0.7_8_0_3, -1.0_6_7_4, -1.2_9_2_0, -0.1_5_7_0, 0.8_0_0_8, 2.0_7_2_3, -0.9_4_9_7, 0.2_7_7_1, -2.2_2_2_7, -0.7_6_1_2, -1.4_3_4_6, -1.2_0_6_1, -1.6_4_2_6, -0.3_0_0_0, -0.7_1_3_9, -1.1_9_3_4, -1.8_6_9_1, -1.6_9_7_3, -1.5_9_4_7, -1.2_7_0_5, -0.3_5_2_3, -0.5_5_1_3] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , a__ , atol=1e-2 , rtol=1e-2 ) @unittest.skip( """Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test""" ) @slow def __A ( self ): _lowerCAmelCase : str = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : Optional[Any] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-70b-hf""" , device_map="""auto""" ) _lowerCAmelCase : str = model(torch.tensor(a__ ) ) _lowerCAmelCase : Tuple = torch.tensor( [[-4.2_3_2_7, -3.3_3_6_0, -4.6_6_6_5, -4.7_6_3_1, -1.8_1_8_0, -3.4_1_7_0, -1.4_2_1_1, -3.1_8_1_0]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , a__ , atol=1e-2 , rtol=1e-2 ) # fmt: off _lowerCAmelCase : Union[str, Any] = torch.tensor([-9.4_9_2_2, -3.9_5_5_1, 1.7_9_9_8, -5.6_7_5_8, -5.1_0_5_5, -5.8_9_8_4, -4.8_3_2_0, -6.8_0_8_6, -6.5_3_9_1, -5.6_1_7_2, -5.5_8_2_0, -5.5_3_5_2, 1.7_8_8_1, 3.6_2_8_9, -6.5_1_1_7, -3.4_7_8_5, -9.5_0_0_0, -6.0_3_5_2, -6.8_1_2_5, -6.0_1_9_5, -6.6_8_3_6, -5.4_7_2_7, -6.2_8_1_2, -6.0_3_9_1, -7.3_3_9_8, -7.4_2_9_7, -7.4_8_4_4, -6.5_8_2_0, -5.8_7_8_9, -5.5_3_1_2] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , a__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("""Model is curently gated""" ) @slow def __A ( self ): _lowerCAmelCase : List[str] = """Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the \"princi""" _lowerCAmelCase : str = """Simply put, the theory of relativity states that """ _lowerCAmelCase : List[str] = LlamaTokenizer.from_pretrained("""meta-llama/Llama-2-13b-chat-hf""" ) _lowerCAmelCase : Optional[Any] = tokenizer.encode(a__ , return_tensors="""pt""" ) _lowerCAmelCase : Optional[Any] = LlamaForCausalLM.from_pretrained( """meta-llama/Llama-2-13b-chat-hf""" , device_map="""sequential""" , use_safetensors=a__ ) # greedy generation outputs _lowerCAmelCase : Optional[int] = model.generate(a__ , max_new_tokens=64 , top_p=a__ , temperature=1 , do_sample=a__ ) _lowerCAmelCase : Dict = tokenizer.decode(generated_ids[0] , skip_special_tokens=a__ ) self.assertEqual(a__ , a__ )
44
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowercase__ ( lowercase ): @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Dict = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : Dict = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : str = '1' _UpperCamelCase : Union[str, Any] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Any = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _UpperCamelCase : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _UpperCamelCase : List[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCamelCase__ ) BertModel.from_pretrained(lowerCamelCase__ ) BertTokenizer.from_pretrained(lowerCamelCase__ ) pipeline(task='fill-mask' ,model=lowerCamelCase__ ) # baseline - just load from_pretrained with normal network _UpperCamelCase : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _UpperCamelCase : List[Any] = self.get_env() _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _UpperCamelCase : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _UpperCamelCase : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Optional[int] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # next emulate no network _UpperCamelCase : Dict = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : Dict = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : int = '\nfrom transformers import pipeline\n ' _UpperCamelCase : str = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _UpperCamelCase : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _UpperCamelCase : Union[str, Any] = self.get_env() _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Tuple = [sys.executable, '-c', '\n'.join([load, mock, run] )] _UpperCamelCase : int = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,1 ,result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' ,result.stderr.decode().replace('\n' ,'' ) ,) @require_torch def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = '\nfrom transformers import AutoModel\n ' _UpperCamelCase : int = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _UpperCamelCase : Any = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _UpperCamelCase : Optional[Any] = self.get_env() _UpperCamelCase : Optional[int] = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _UpperCamelCase : List[Any] = '1' _UpperCamelCase : Dict = subprocess.run(lowerCamelCase__ ,env=lowerCamelCase__ ,check=lowerCamelCase__ ,capture_output=lowerCamelCase__ ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() )
83
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase_ = { "configuration_mask2former": [ "MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "Mask2FormerConfig", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["Mask2FormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "Mask2FormerForUniversalSegmentation", "Mask2FormerModel", "Mask2FormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure)
45
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase__ ( unittest.TestCase ): def __init__( self : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=99 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=512 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : int=0.0_2 ,lowerCamelCase__ : int=4 ,): '''simple docstring''' _UpperCamelCase : List[Any] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Optional[Any] = is_training _UpperCamelCase : Optional[int] = use_attention_mask _UpperCamelCase : Any = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : str = intermediate_size _UpperCamelCase : int = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Optional[int] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_choices def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Union[str, Any] = None if self.use_attention_mask: _UpperCamelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=lowerCamelCase__ ,) return config, input_ids, attention_mask def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Optional[int] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase : Dict = model_class_name.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase : Tuple = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] _UpperCamelCase : Any = (1, 11, 768) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1E-4 ) )
83
0
"""simple docstring""" def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str = " " ): '''simple docstring''' lowerCAmelCase = [] lowerCAmelCase = 0 for index, char in enumerate(SCREAMING_SNAKE_CASE ): if char == separator: split_words.append(string[last_index:index] ) lowerCAmelCase = index + 1 elif index + 1 == len(SCREAMING_SNAKE_CASE ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
46
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer snake_case_ : List[Any] = logging.get_logger(__name__) class lowercase__ ( lowercase ): lowercase__ = """AutoTokenizer""" lowercase__ = ["""tokenizer"""] lowercase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self : List[str] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Tuple=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ) _UpperCamelCase : Dict = speaker_embeddings @classmethod def UpperCamelCase_ ( cls : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : str="speaker_embeddings_path.json" ,**lowerCamelCase__ : Optional[Any] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: _UpperCamelCase : Optional[Any] = get_file_from_repo( lowerCamelCase__ ,lowerCamelCase__ ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(lowerCamelCase__ ,lowerCamelCase__ )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) _UpperCamelCase : Union[str, Any] = None else: with open(lowerCamelCase__ ) as speaker_embeddings_json: _UpperCamelCase : Optional[int] = json.load(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = None _UpperCamelCase : Tuple = AutoTokenizer.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) return cls(tokenizer=lowerCamelCase__ ,speaker_embeddings=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int="speaker_embeddings_path.json" ,lowerCamelCase__ : Dict="speaker_embeddings" ,lowerCamelCase__ : bool = False ,**lowerCamelCase__ : Tuple ,): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ,'v2' ) ,exist_ok=lowerCamelCase__ ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Optional[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCamelCase : Any = self._load_voice_preset(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] ,lowerCamelCase__ ,F'{prompt_key}_{key}' ) ,voice_preset[key] ,allow_pickle=lowerCamelCase__ ,) _UpperCamelCase : List[str] = os.path.join(lowerCamelCase__ ,F'{prompt_key}_{key}.npy' ) _UpperCamelCase : str = tmp_dict with open(os.path.join(lowerCamelCase__ ,lowerCamelCase__ ) ,'w' ) as fp: json.dump(lowerCamelCase__ ,lowerCamelCase__ ) super().save_pretrained(lowerCamelCase__ ,lowerCamelCase__ ,**lowerCamelCase__ ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : str = None ,**lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.speaker_embeddings[voice_preset] _UpperCamelCase : Union[str, Any] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) _UpperCamelCase : Dict = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' ,'/' ) ,voice_preset_paths[key] ,subfolder=kwargs.pop('subfolder' ,lowerCamelCase__ ) ,cache_dir=kwargs.pop('cache_dir' ,lowerCamelCase__ ) ,force_download=kwargs.pop('force_download' ,lowerCamelCase__ ) ,proxies=kwargs.pop('proxies' ,lowerCamelCase__ ) ,resume_download=kwargs.pop('resume_download' ,lowerCamelCase__ ) ,local_files_only=kwargs.pop('local_files_only' ,lowerCamelCase__ ) ,use_auth_token=kwargs.pop('use_auth_token' ,lowerCamelCase__ ) ,revision=kwargs.pop('revision' ,lowerCamelCase__ ) ,) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" ,"/" ) ,voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) _UpperCamelCase : List[str] = np.load(lowerCamelCase__ ) return voice_preset_dict def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] ,np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self : Any ,lowerCamelCase__ : Optional[Any]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Any="pt" ,lowerCamelCase__ : Dict=256 ,lowerCamelCase__ : int=False ,lowerCamelCase__ : int=True ,lowerCamelCase__ : List[str]=False ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' if voice_preset is not None and not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): if ( isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCamelCase : Optional[int] = self._load_voice_preset(lowerCamelCase__ ) else: if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) and not voice_preset.endswith('.npz' ): _UpperCamelCase : Tuple = voice_preset + '.npz' _UpperCamelCase : str = np.load(lowerCamelCase__ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = BatchFeature(data=lowerCamelCase__ ,tensor_type=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.tokenizer( lowerCamelCase__ ,return_tensors=lowerCamelCase__ ,padding='max_length' ,max_length=lowerCamelCase__ ,return_attention_mask=lowerCamelCase__ ,return_token_type_ids=lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ,**lowerCamelCase__ ,) if voice_preset is not None: _UpperCamelCase : Optional[Any] = voice_preset return encoded_text
83
0
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCamelCase : Tuple = "pt" elif is_tf_available(): lowerCamelCase : Optional[int] = "tf" else: lowerCamelCase : Tuple = "jax" class A__ ( A__ , unittest.TestCase ): A__ = ByTaTokenizer A__ = False def A ( self : str ) -> Any: '''simple docstring''' super().setUp() _SCREAMING_SNAKE_CASE =ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def A ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' return ByTaTokenizer.from_pretrained('google/byt5-small' ) def A ( self : Dict , **_a : Union[str, Any] ) -> ByTaTokenizer: '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname , **_a ) def A ( self : List[str] , _a : Dict , _a : str=False , _a : Optional[int]=20 , _a : int=5 ) -> Tuple[str, list]: '''simple docstring''' _SCREAMING_SNAKE_CASE =[] for i in range(len(_a ) ): try: _SCREAMING_SNAKE_CASE =tokenizer.decode([i] , clean_up_tokenization_spaces=_a ) except UnicodeDecodeError: pass toks.append((i, tok) ) _SCREAMING_SNAKE_CASE =list(filter(lambda _a : re.match(r'^[ a-zA-Z]+$' , t[1] ) , _a ) ) _SCREAMING_SNAKE_CASE =list(filter(lambda _a : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_a ) , _a ) ) if max_length is not None and len(_a ) > max_length: _SCREAMING_SNAKE_CASE =toks[:max_length] if min_length is not None and len(_a ) < min_length and len(_a ) > 0: while len(_a ) < min_length: _SCREAMING_SNAKE_CASE =toks + toks # toks_str = [t[1] for t in toks] _SCREAMING_SNAKE_CASE =[t[0] for t in toks] # Ensure consistency _SCREAMING_SNAKE_CASE =tokenizer.decode(_a , clean_up_tokenization_spaces=_a ) if " " not in output_txt and len(_a ) > 1: _SCREAMING_SNAKE_CASE =( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_a ) + ' ' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_a ) ) if with_prefix_space: _SCREAMING_SNAKE_CASE =' ' + output_txt _SCREAMING_SNAKE_CASE =tokenizer.encode(_a , add_special_tokens=_a ) return output_txt, output_ids def A ( self : Any ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE =tokenizer(['hi</s>', 'I went to the gym</s>', '</s>'] ) _SCREAMING_SNAKE_CASE =tokenizer(['hi', 'I went to the gym', ''] ) self.assertListEqual(batch_with_eos_added['input_ids'] , batch_without_eos_added['input_ids'] ) def A ( self : Optional[int] ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE ='Unicode €.' _SCREAMING_SNAKE_CASE =tokenizer(_a ) _SCREAMING_SNAKE_CASE =[88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['input_ids'] , _a ) # decoding _SCREAMING_SNAKE_CASE =tokenizer.decode(_a ) self.assertEqual(_a , 'Unicode €.</s>' ) _SCREAMING_SNAKE_CASE =tokenizer('e è é ê ë' ) _SCREAMING_SNAKE_CASE =[104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['input_ids'] , _a ) # decoding _SCREAMING_SNAKE_CASE =tokenizer.decode(_a ) self.assertEqual(_a , 'e è é ê ë</s>' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ) , 'e è é ê ë</s>' ) def A ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE =['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off _SCREAMING_SNAKE_CASE =[68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on _SCREAMING_SNAKE_CASE =tokenizer(_a , padding=_a , return_tensors=_a ) self.assertIsInstance(_a , _a ) if FRAMEWORK != "jax": _SCREAMING_SNAKE_CASE =list(batch.input_ids.numpy()[0] ) else: _SCREAMING_SNAKE_CASE =list(batch.input_ids.tolist()[0] ) self.assertListEqual(_a , _a ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def A ( self : Dict ) -> Union[str, Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE =['A long paragraph for summarization.', 'Another paragraph for summarization.'] _SCREAMING_SNAKE_CASE =tokenizer(_a , padding=_a , return_tensors=_a ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids' , _a ) self.assertIn('attention_mask' , _a ) self.assertNotIn('decoder_input_ids' , _a ) self.assertNotIn('decoder_attention_mask' , _a ) def A ( self : List[str] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE =[ 'Summary of the text.', 'Another summary.', ] _SCREAMING_SNAKE_CASE =tokenizer( text_target=_a , max_length=32 , padding='max_length' , truncation=_a , return_tensors=_a ) self.assertEqual(32 , targets['input_ids'].shape[1] ) def A ( self : Union[str, Any] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.ta_base_tokenizer _SCREAMING_SNAKE_CASE =['A long paragraph for summarization. </s>'] _SCREAMING_SNAKE_CASE =['Summary of the text. </s>'] # fmt: off _SCREAMING_SNAKE_CASE =[68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] _SCREAMING_SNAKE_CASE =[86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on _SCREAMING_SNAKE_CASE =tokenizer(_a , text_target=_a ) self.assertEqual(_a , batch['input_ids'][0] ) self.assertEqual(_a , batch['labels'][0] ) def A ( self : Tuple ) -> List[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test _SCREAMING_SNAKE_CASE =self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc _SCREAMING_SNAKE_CASE =tempfile.mkdtemp() _SCREAMING_SNAKE_CASE =' He is very happy, UNwant\u00E9d,running' _SCREAMING_SNAKE_CASE =tokenizer.encode(_a , add_special_tokens=_a ) tokenizer.save_pretrained(_a ) _SCREAMING_SNAKE_CASE =tokenizer.__class__.from_pretrained(_a ) _SCREAMING_SNAKE_CASE =after_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) shutil.rmtree(_a ) _SCREAMING_SNAKE_CASE =self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc _SCREAMING_SNAKE_CASE =tempfile.mkdtemp() _SCREAMING_SNAKE_CASE =' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) _SCREAMING_SNAKE_CASE =tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) _SCREAMING_SNAKE_CASE =tokenizer.encode(_a , add_special_tokens=_a ) tokenizer.save_pretrained(_a ) _SCREAMING_SNAKE_CASE =tokenizer.__class__.from_pretrained(_a ) _SCREAMING_SNAKE_CASE =after_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) self.assertIn('new_additional_special_token' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) _SCREAMING_SNAKE_CASE =tokenizer.__class__.from_pretrained(_a , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_a ) def A ( self : int ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =[] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_a ) with open(os.path.join(_a , 'special_tokens_map.json' ) , encoding='utf-8' ) as json_file: _SCREAMING_SNAKE_CASE =json.load(_a ) with open(os.path.join(_a , 'tokenizer_config.json' ) , encoding='utf-8' ) as json_file: _SCREAMING_SNAKE_CASE =json.load(_a ) _SCREAMING_SNAKE_CASE =[f"<extra_id_{i}>" for i in range(125 )] _SCREAMING_SNAKE_CASE =added_tokens_extra_ids + [ 'an_additional_special_token' ] _SCREAMING_SNAKE_CASE =added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(_a , 'special_tokens_map.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(_a , _a ) with open(os.path.join(_a , 'tokenizer_config.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(_a , _a ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files _SCREAMING_SNAKE_CASE =tokenizer_class.from_pretrained( _a , ) self.assertIn( 'an_additional_special_token' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['an_additional_special_token'] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained _SCREAMING_SNAKE_CASE =added_tokens_extra_ids + [AddedToken('a_new_additional_special_token' , lstrip=_a )] _SCREAMING_SNAKE_CASE =tokenizer_class.from_pretrained( _a , additional_special_tokens=_a , ) self.assertIn('a_new_additional_special_token' , tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ) , ) def A ( self : Union[str, Any] ) -> int: '''simple docstring''' _SCREAMING_SNAKE_CASE =[] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_a ) _SCREAMING_SNAKE_CASE =tokenizer_class.from_pretrained(_a ) self.assertTrue(tokenizer.decode([255] ) == '' ) def A ( self : int ) -> List[Any]: '''simple docstring''' pass def A ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def A ( self : List[Any] ) -> Tuple: '''simple docstring''' pass def A ( self : Tuple ) -> Optional[Any]: '''simple docstring''' pass def A ( self : Optional[int] ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.get_tokenizers(fast=_a , do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): _SCREAMING_SNAKE_CASE =['t', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 'x', 't', '</s>'] _SCREAMING_SNAKE_CASE =tokenizer.convert_tokens_to_string(_a ) self.assertIsInstance(_a , _a ) def A ( self : List[str] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): _SCREAMING_SNAKE_CASE =[ 'bos_token', 'eos_token', 'unk_token', 'sep_token', 'pad_token', 'cls_token', 'mask_token', ] _SCREAMING_SNAKE_CASE =0 _SCREAMING_SNAKE_CASE =tokenizer.convert_ids_to_tokens( _a , skip_special_tokens=_a ) for attr in attributes_list: setattr(_a , attr + '_id' , _a ) self.assertEqual(getattr(_a , _a ) , _a ) self.assertEqual(getattr(_a , attr + '_id' ) , _a ) setattr(_a , attr + '_id' , _a ) self.assertEqual(getattr(_a , _a ) , _a ) self.assertEqual(getattr(_a , attr + '_id' ) , _a ) setattr(_a , 'additional_special_tokens_ids' , [] ) self.assertListEqual(getattr(_a , 'additional_special_tokens' ) , [] ) self.assertListEqual(getattr(_a , 'additional_special_tokens_ids' ) , [] ) setattr(_a , 'additional_special_tokens_ids' , [token_id_to_test_setters] ) self.assertListEqual(getattr(_a , 'additional_special_tokens' ) , [token_to_test_setters] ) self.assertListEqual(getattr(_a , 'additional_special_tokens_ids' ) , [token_id_to_test_setters] )
47
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin snake_case_ : Tuple = random.Random() def A__ ( UpperCAmelCase_ , UpperCAmelCase_=1.0 , UpperCAmelCase_=None , UpperCAmelCase_=None ): if rng is None: _UpperCamelCase : Dict = global_rng _UpperCamelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int=7 ,lowerCamelCase__ : str=400 ,lowerCamelCase__ : int=2000 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=16000 ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Optional[int]=True ,): '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Union[str, Any] = batch_size _UpperCamelCase : List[str] = min_seq_length _UpperCamelCase : Optional[int] = max_seq_length _UpperCamelCase : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _UpperCamelCase : List[str] = feature_size _UpperCamelCase : List[str] = padding_value _UpperCamelCase : List[Any] = sampling_rate _UpperCamelCase : Dict = return_attention_mask _UpperCamelCase : Tuple = do_normalize def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : Tuple=False ): '''simple docstring''' def _flatten(lowerCamelCase__ : Optional[Any] ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _UpperCamelCase : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _UpperCamelCase : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length ,self.max_seq_length ,self.seq_length_diff ) ] if numpify: _UpperCamelCase : int = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = WavaVecaFeatureExtractor def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : List[str] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): '''simple docstring''' self.assertTrue(np.all(np.mean(lowerCamelCase__ ,axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ,axis=0 ) - 1 ) < 1E-3 ) ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' # Tests that all call wrap to encode_plus and batch_encode_plus _UpperCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _UpperCamelCase : int = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Tuple = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test not batched input _UpperCamelCase : Tuple = feat_extract(speech_inputs[0] ,return_tensors='np' ).input_values _UpperCamelCase : Any = feat_extract(np_speech_inputs[0] ,return_tensors='np' ).input_values self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test batched _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : Optional[int] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _UpperCamelCase : str = [floats_list((1, x) )[0] for x in (800, 800, 800)] _UpperCamelCase : str = np.asarray(lowerCamelCase__ ) _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values _UpperCamelCase : int = feat_extract(lowerCamelCase__ ,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ ,lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ ,lowerCamelCase__ ,atol=1E-3 ) ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : str = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : List[str] = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = feat_extract(lowerCamelCase__ ,padding=lowerCamelCase__ ,max_length=lowerCamelCase__ ,return_tensors='np' ) _UpperCamelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[str] = range(800 ,1400 ,200 ) _UpperCamelCase : List[str] = [floats_list((1, x) )[0] for x in lengths] _UpperCamelCase : Optional[Any] = ['longest', 'max_length', 'do_not_pad'] _UpperCamelCase : str = [None, 1600, None] for max_length, padding in zip(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : List[str] = feat_extract(lowerCamelCase__ ,max_length=lowerCamelCase__ ,padding=lowerCamelCase__ ) _UpperCamelCase : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Union[str, Any] = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='max_length' ,return_tensors='np' ) _UpperCamelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : int = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=1000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) _UpperCamelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(800 ,1400 ,200 )] _UpperCamelCase : Any = feat_extract( lowerCamelCase__ ,truncation=lowerCamelCase__ ,max_length=2000 ,padding='longest' ,return_tensors='np' ) _UpperCamelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase_ ( self : Any ): '''simple docstring''' import torch _UpperCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _UpperCamelCase : Optional[int] = np.random.rand(100 ).astype(np.floataa ) _UpperCamelCase : Dict = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _UpperCamelCase : Optional[int] = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _UpperCamelCase : Tuple = feature_extractor.pad([{'input_values': inputs}] ,return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: _UpperCamelCase : Optional[int] = WavaVecaConfig.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase__ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask ,config.feat_extract_norm == 'layer' )
83
0
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: lowerCamelCase : Optional[Any] = 1.5 lowerCamelCase : Optional[int] = int(factor * num_class_images ) lowerCamelCase : str = ClipClient( url="https://knn.laion.ai/knn-service" ,indice_name="laion_400m" ,num_images=_SCREAMING_SNAKE_CASE ,aesthetic_weight=0.1 ) os.makedirs(f'''{class_data_dir}/images''' ,exist_ok=_SCREAMING_SNAKE_CASE ) if len(list(Path(f'''{class_data_dir}/images''' ).iterdir() ) ) >= num_class_images: return while True: lowerCamelCase : List[Any] = client.query(text=_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) >= factor * num_class_images or num_images > 1e4: break else: lowerCamelCase : Any = int(factor * num_images ) lowerCamelCase : int = ClipClient( url="https://knn.laion.ai/knn-service" ,indice_name="laion_400m" ,num_images=_SCREAMING_SNAKE_CASE ,aesthetic_weight=0.1 ,) lowerCamelCase : Any = 0 lowerCamelCase : str = 0 lowerCamelCase : Dict = tqdm(desc="downloading real regularization images" ,total=_SCREAMING_SNAKE_CASE ) with open(f'''{class_data_dir}/caption.txt''' ,"w" ) as fa, open(f'''{class_data_dir}/urls.txt''' ,"w" ) as fa, open( f'''{class_data_dir}/images.txt''' ,"w" ) as fa: while total < num_class_images: lowerCamelCase : str = class_images[count] count += 1 try: lowerCamelCase : Union[str, Any] = requests.get(images["url"] ) if img.status_code == 200: lowerCamelCase : Dict = Image.open(BytesIO(img.content ) ) with open(f'''{class_data_dir}/images/{total}.jpg''' ,"wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'''{class_data_dir}/images/{total}.jpg''' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def A ( ) -> List[Any]: lowerCamelCase : Dict = argparse.ArgumentParser("" ,add_help=_SCREAMING_SNAKE_CASE ) parser.add_argument("--class_prompt" ,help="text prompt to retrieve images" ,required=_SCREAMING_SNAKE_CASE ,type=_SCREAMING_SNAKE_CASE ) parser.add_argument("--class_data_dir" ,help="path to save images" ,required=_SCREAMING_SNAKE_CASE ,type=_SCREAMING_SNAKE_CASE ) parser.add_argument("--num_class_images" ,help="number of images to download" ,default=200 ,type=_SCREAMING_SNAKE_CASE ) return parser.parse_args() if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[str] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
48
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1 , UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : int = 1 _UpperCamelCase : Union[str, Any] = 0 for divide_by_number in range(UpperCAmelCase_ , digit + 1 ): _UpperCamelCase : list[int] = [] _UpperCamelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = len(UpperCAmelCase_ ) _UpperCamelCase : List[Any] = divide_by_number else: has_been_divided.append(UpperCAmelCase_ ) _UpperCamelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
83
0
import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): # Initialise PyTorch model __a = BigBirdConfig.from_json_file(_UpperCAmelCase ) print(f'Building PyTorch model from configuration: {config}' ) if is_trivia_qa: __a = BigBirdForQuestionAnswering(_UpperCAmelCase ) else: __a = BigBirdForPreTraining(_UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(_UpperCAmelCase , _UpperCAmelCase , is_trivia_qa=_UpperCAmelCase ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __snake_case :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--big_bird_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_trivia_qa''', action='''store_true''', help='''Whether to convert a model with a trivia_qa head.''' ) __snake_case :Any = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
49
'''simple docstring''' def A__ ( UpperCAmelCase_ ): if num < 0: return False _UpperCamelCase : int = num _UpperCamelCase : int = 0 while num > 0: _UpperCamelCase : str = rev_num * 1_0 + (num % 1_0) num //= 1_0 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
83
0
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ) -> Union[str, Any]: if isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Tuple = len(set_a.intersection(_UpperCAmelCase ) ) if alternative_union: lowerCamelCase__ : Union[str, Any] = len(_UpperCAmelCase ) + len(_UpperCAmelCase ) else: lowerCamelCase__ : Any = len(set_a.union(_UpperCAmelCase ) ) return intersection / union if isinstance(_UpperCAmelCase , (list, tuple) ) and isinstance(_UpperCAmelCase , (list, tuple) ): lowerCamelCase__ : int = [element for element in set_a if element in set_b] if alternative_union: lowerCamelCase__ : int = len(_UpperCAmelCase ) + len(_UpperCAmelCase ) return len(_UpperCAmelCase ) / union else: lowerCamelCase__ : Dict = set_a + [element for element in set_b if element not in set_a] return len(_UpperCAmelCase ) / len(_UpperCAmelCase ) return len(_UpperCAmelCase ) / len(_UpperCAmelCase ) return None if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = {"""a""", """b""", """c""", """d""", """e"""} _UpperCAmelCase : Tuple = {"""c""", """d""", """e""", """f""", """h""", """i"""} print(jaccard_similarity(set_a, set_b))
50
'''simple docstring''' def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[str] = abs(UpperCAmelCase_ ) _UpperCamelCase : int = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def A__ ( UpperCAmelCase_ ): _UpperCamelCase : List[Any] = abs(UpperCAmelCase_ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def A__ ( UpperCAmelCase_ ): return sum(int(UpperCAmelCase_ ) for c in str(abs(UpperCAmelCase_ ) ) ) def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) -> None: _UpperCamelCase : str = f'{func.__name__}({value})' _UpperCamelCase : Tuple = timeit(f'__main__.{call}' , setup='import __main__' ) print(f'{call:56} = {func(UpperCAmelCase_ )} -- {timing:.4f} seconds' ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCAmelCase_ , UpperCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
83
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable snake_case_ : Union[str, Any] = {"configuration_gpt_neox": ["GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoXConfig"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = ["GPTNeoXTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : str = [ "GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoXForCausalLM", "GPTNeoXForQuestionAnswering", "GPTNeoXForSequenceClassification", "GPTNeoXForTokenClassification", "GPTNeoXLayer", "GPTNeoXModel", "GPTNeoXPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys snake_case_ : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
51
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
0
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class A__ : def __init__( self , A_ = None ): '''simple docstring''' if components is None: UpperCamelCase : Any = [] UpperCamelCase : List[str] = list(A_ ) def __len__( self ): '''simple docstring''' return len(self.__components ) def __str__( self ): '''simple docstring''' return "(" + ",".join(map(A_ , self.__components ) ) + ")" def __add__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = len(self ) if size == len(A_ ): UpperCamelCase : Optional[Any] = [self.__components[i] + other.component(A_ ) for i in range(A_ )] return Vector(A_ ) else: raise Exception("must have the same size" ) def __sub__( self , A_ ): '''simple docstring''' UpperCamelCase : Tuple = len(self ) if size == len(A_ ): UpperCamelCase : Tuple = [self.__components[i] - other.component(A_ ) for i in range(A_ )] return Vector(A_ ) else: # error case raise Exception("must have the same size" ) @overload def __mul__( self , A_ ): '''simple docstring''' ... @overload def __mul__( self , A_ ): '''simple docstring''' ... def __mul__( self , A_ ): '''simple docstring''' if isinstance(A_ , (float, int) ): UpperCamelCase : Optional[Any] = [c * other for c in self.__components] return Vector(A_ ) elif isinstance(A_ , A_ ) and len(self ) == len(A_ ): UpperCamelCase : Tuple = len(self ) UpperCamelCase : Optional[int] = [self.__components[i] * other.component(A_ ) for i in range(A_ )] return sum(A_ ) else: # error case raise Exception("invalid operand!" ) def __UpperCamelCase( self ): '''simple docstring''' return Vector(self.__components ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if isinstance(A_ , A_ ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception("index out of range" ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' assert -len(self.__components ) <= pos < len(self.__components ) UpperCamelCase : Any = value def __UpperCamelCase( self ): '''simple docstring''' if len(self.__components ) == 0: raise Exception("Vector is empty" ) UpperCamelCase : Dict = [c**2 for c in self.__components] return math.sqrt(sum(A_ ) ) def __UpperCamelCase( self , A_ , A_ = False ): '''simple docstring''' UpperCamelCase : int = self * other UpperCamelCase : str = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def A_ ( _lowerCAmelCase ) -> Vector: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) return Vector([0] * dimension ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Vector: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (isinstance(_lowerCAmelCase , _lowerCAmelCase )) UpperCamelCase : Dict = [0] * dimension UpperCamelCase : str = 1 return Vector(_lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Vector: assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (isinstance(_lowerCAmelCase , (int, float) )) ) return x * scalar + y def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Vector: random.seed(_lowerCAmelCase ) UpperCamelCase : str = [random.randint(_lowerCAmelCase , _lowerCAmelCase ) for _ in range(_lowerCAmelCase )] return Vector(_lowerCAmelCase ) class A__ : def __init__( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = matrix UpperCamelCase : List[str] = w UpperCamelCase : Any = h def __str__( self ): '''simple docstring''' UpperCamelCase : Dict = "" for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , A_ ): '''simple docstring''' if self.__width == other.width() and self.__height == other.height(): UpperCamelCase : Optional[Any] = [] for i in range(self.__height ): UpperCamelCase : Optional[int] = [ self.__matrix[i][j] + other.component(A_ , A_ ) for j in range(self.__width ) ] matrix.append(A_ ) return Matrix(A_ , self.__width , self.__height ) else: raise Exception("matrix must have the same dimension!" ) def __sub__( self , A_ ): '''simple docstring''' if self.__width == other.width() and self.__height == other.height(): UpperCamelCase : List[Any] = [] for i in range(self.__height ): UpperCamelCase : Dict = [ self.__matrix[i][j] - other.component(A_ , A_ ) for j in range(self.__width ) ] matrix.append(A_ ) return Matrix(A_ , self.__width , self.__height ) else: raise Exception("matrices must have the same dimension!" ) @overload def __mul__( self , A_ ): '''simple docstring''' ... @overload def __mul__( self , A_ ): '''simple docstring''' ... def __mul__( self , A_ ): '''simple docstring''' if isinstance(A_ , A_ ): # matrix-vector if len(A_ ) == self.__width: UpperCamelCase : int = zero_vector(self.__height ) for i in range(self.__height ): UpperCamelCase : int = [ self.__matrix[i][j] * other.component(A_ ) for j in range(self.__width ) ] ans.change_component(A_ , sum(A_ ) ) return ans else: raise Exception( "vector must have the same size as the " "number of columns of the matrix!" ) elif isinstance(A_ , (int, float) ): # matrix-scalar UpperCamelCase : Dict = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(A_ , self.__width , self.__height ) return None def __UpperCamelCase( self ): '''simple docstring''' return self.__height def __UpperCamelCase( self ): '''simple docstring''' return self.__width def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception("change_component: indices out of bounds" ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' if 0 <= x < self.__height and 0 <= y < self.__width: UpperCamelCase : List[str] = value else: raise Exception("change_component: indices out of bounds" ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if self.__height != self.__width: raise Exception("Matrix is not square" ) UpperCamelCase : List[Any] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(A_ ) ): UpperCamelCase : int = minor[i][:y] + minor[i][y + 1 :] return Matrix(A_ , self.__width - 1 , self.__height - 1 ).determinant() def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if self.__height != self.__width: raise Exception("Matrix is not square" ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(A_ , A_ ) else: raise Exception("Indices out of bounds" ) def __UpperCamelCase( self ): '''simple docstring''' if self.__height != self.__width: raise Exception("Matrix is not square" ) if self.__height < 1: raise Exception("Matrix has no element" ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: UpperCamelCase : Union[str, Any] = [ self.__matrix[0][y] * self.cofactor(0 , A_ ) for y in range(self.__width ) ] return sum(A_ ) def A_ ( _lowerCAmelCase ) -> Matrix: UpperCamelCase : list[list[float]] = [[0] * n for _ in range(_lowerCAmelCase )] return Matrix(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Matrix: random.seed(_lowerCAmelCase ) UpperCamelCase : list[list[float]] = [ [random.randint(_lowerCAmelCase , _lowerCAmelCase ) for _ in range(_lowerCAmelCase )] for _ in range(_lowerCAmelCase ) ] return Matrix(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase )
52
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : Optional[Any] = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowercase__ ( lowercase ): lowercase__ = """mvp""" lowercase__ = ["""past_key_values"""] lowercase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] ,lowerCamelCase__ : Any=50267 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : int=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Union[str, Any]=16 ,lowerCamelCase__ : List[Any]=12 ,lowerCamelCase__ : Tuple=4096 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : Optional[int]=0.0 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Tuple=0.1 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Union[str, Any]=0.0_2 ,lowerCamelCase__ : Union[str, Any]=0.0 ,lowerCamelCase__ : Tuple=False ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : str=1 ,lowerCamelCase__ : Any=0 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=2 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=False ,lowerCamelCase__ : Tuple=100 ,lowerCamelCase__ : Optional[int]=800 ,**lowerCamelCase__ : int ,): '''simple docstring''' _UpperCamelCase : Optional[int] = vocab_size _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Dict = d_model _UpperCamelCase : Any = encoder_ffn_dim _UpperCamelCase : Dict = encoder_layers _UpperCamelCase : Optional[Any] = encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : str = decoder_layers _UpperCamelCase : int = decoder_attention_heads _UpperCamelCase : str = dropout _UpperCamelCase : str = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : Dict = activation_function _UpperCamelCase : List[str] = init_std _UpperCamelCase : Dict = encoder_layerdrop _UpperCamelCase : Tuple = decoder_layerdrop _UpperCamelCase : Optional[int] = classifier_dropout _UpperCamelCase : str = use_cache _UpperCamelCase : Union[str, Any] = encoder_layers _UpperCamelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : Any = use_prompt _UpperCamelCase : Optional[int] = prompt_length _UpperCamelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,forced_eos_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' ,lowerCamelCase__ ): _UpperCamelCase : Union[str, Any] = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
83
0
'''simple docstring''' import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class snake_case ( __lowerCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict =RoCBertTokenizer SCREAMING_SNAKE_CASE_ : int =None SCREAMING_SNAKE_CASE_ : Any =False SCREAMING_SNAKE_CASE_ : int =True SCREAMING_SNAKE_CASE_ : Optional[Any] =filter_non_english def _lowerCamelCase ( self : int ): super().setUp() __UpperCamelCase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', '你', '好', '是', '谁', 'a', 'b', 'c', 'd'] __UpperCamelCase = {} __UpperCamelCase = {} for i, value in enumerate(__A ): __UpperCamelCase = i __UpperCamelCase = i __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['word_shape_file'] ) __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['word_pronunciation_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.word_shape_file , 'w' , encoding='utf-8' ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , 'w' , encoding='utf-8' ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __UpperCamelCase = tokenizer.tokenize('你好[SEP]你是谁' ) self.assertListEqual(__A , ['你', '好', '[SEP]', '你', '是', '谁'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def _lowerCamelCase ( self : int ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def _lowerCamelCase ( self : int ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = RoCBertBasicTokenizer(do_lower_case=__A , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] __UpperCamelCase = {} for i, token in enumerate(__A ): __UpperCamelCase = i __UpperCamelCase = RoCBertWordpieceTokenizer(vocab=__A , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def _lowerCamelCase ( self : Dict ): self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def _lowerCamelCase ( self : Dict ): self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def _lowerCamelCase ( self : int ): self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) if self.test_rust_tokenizer: __UpperCamelCase = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) def _lowerCamelCase ( self : Dict ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__A , **__A ) __UpperCamelCase = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' __UpperCamelCase = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) __UpperCamelCase = tokenizer_r.do_lower_case if hasattr(__A , 'do_lower_case' ) else False __UpperCamelCase = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), 'Allen'), ((2_1, 2_3), '##NL'), ((2_3, 2_4), '##P'), ((2_5, 3_3), 'sentence'), ((3_3, 3_4), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), 'allen'), ((2_1, 2_3), '##nl'), ((2_3, 2_4), '##p'), ((2_5, 3_3), 'sentence'), ((3_3, 3_4), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = ['的', '人', '有'] __UpperCamelCase = ''.join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCamelCase = True __UpperCamelCase = self.tokenizer_class.from_pretrained(__A , **__A ) __UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__A , **__A ) __UpperCamelCase = tokenizer_p.encode(__A , add_special_tokens=__A ) __UpperCamelCase = tokenizer_r.encode(__A , add_special_tokens=__A ) __UpperCamelCase = tokenizer_r.convert_ids_to_tokens(__A ) __UpperCamelCase = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) __UpperCamelCase = False __UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__A , **__A ) __UpperCamelCase = self.tokenizer_class.from_pretrained(__A , **__A ) __UpperCamelCase = tokenizer_r.encode(__A , add_special_tokens=__A ) __UpperCamelCase = tokenizer_p.encode(__A , add_special_tokens=__A ) __UpperCamelCase = tokenizer_r.convert_ids_to_tokens(__A ) __UpperCamelCase = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". __UpperCamelCase = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __UpperCamelCase = tokenizer.encode('你好' , add_special_tokens=__A ) __UpperCamelCase = tokenizer.encode('你是谁' , add_special_tokens=__A ) __UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__A ) __UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCamelCase = '你好,你是谁' __UpperCamelCase = tokenizer.tokenize(__A ) __UpperCamelCase = tokenizer.convert_tokens_to_ids(__A ) __UpperCamelCase = tokenizer.convert_tokens_to_shape_ids(__A ) __UpperCamelCase = tokenizer.convert_tokens_to_pronunciation_ids(__A ) __UpperCamelCase = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) __UpperCamelCase = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
53
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
0
"""simple docstring""" import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class UpperCamelCase_ ( tf.keras.optimizers.schedules.LearningRateSchedule): """simple docstring""" def __init__( self : List[Any] , UpperCAmelCase__ : float , UpperCAmelCase__ : Callable , UpperCAmelCase__ : int , UpperCAmelCase__ : float = 1.0 , UpperCAmelCase__ : str = None , ) -> Tuple: super().__init__() __SCREAMING_SNAKE_CASE = initial_learning_rate __SCREAMING_SNAKE_CASE = warmup_steps __SCREAMING_SNAKE_CASE = power __SCREAMING_SNAKE_CASE = decay_schedule_fn __SCREAMING_SNAKE_CASE = name def __call__( self : List[str] , UpperCAmelCase__ : List[Any] ) -> Any: with tf.name_scope(self.name or "WarmUp" ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. __SCREAMING_SNAKE_CASE = tf.cast(UpperCAmelCase__ , tf.floataa ) __SCREAMING_SNAKE_CASE = tf.cast(self.warmup_steps , tf.floataa ) __SCREAMING_SNAKE_CASE = global_step_float / warmup_steps_float __SCREAMING_SNAKE_CASE = self.initial_learning_rate * tf.math.pow(UpperCAmelCase__ , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=UpperCAmelCase__ , ) def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 0.9 , lowerCAmelCase_ = 0.999 , lowerCAmelCase_ = 1E-8 , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 1.0 , lowerCAmelCase_ = None , ): '''simple docstring''' __SCREAMING_SNAKE_CASE = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=lowerCAmelCase_ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=lowerCAmelCase_ , ) if num_warmup_steps: __SCREAMING_SNAKE_CASE = WarmUp( initial_learning_rate=lowerCAmelCase_ , decay_schedule_fn=lowerCAmelCase_ , warmup_steps=lowerCAmelCase_ , ) if weight_decay_rate > 0.0: __SCREAMING_SNAKE_CASE = AdamWeightDecay( learning_rate=lowerCAmelCase_ , weight_decay_rate=lowerCAmelCase_ , beta_a=lowerCAmelCase_ , beta_a=lowerCAmelCase_ , epsilon=lowerCAmelCase_ , clipnorm=lowerCAmelCase_ , global_clipnorm=lowerCAmelCase_ , exclude_from_weight_decay=["LayerNorm", "layer_norm", "bias"] , include_in_weight_decay=lowerCAmelCase_ , ) else: __SCREAMING_SNAKE_CASE = tf.keras.optimizers.Adam( learning_rate=lowerCAmelCase_ , beta_a=lowerCAmelCase_ , beta_a=lowerCAmelCase_ , epsilon=lowerCAmelCase_ , clipnorm=lowerCAmelCase_ , global_clipnorm=lowerCAmelCase_ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Tuple , UpperCAmelCase__ : Union[float, tf.keras.optimizers.schedules.LearningRateSchedule] = 0.001 , UpperCAmelCase__ : float = 0.9 , UpperCAmelCase__ : float = 0.999 , UpperCAmelCase__ : float = 1E-7 , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : Optional[List[str]] = None , UpperCAmelCase__ : Optional[List[str]] = None , UpperCAmelCase__ : str = "AdamWeightDecay" , **UpperCAmelCase__ : Optional[Any] , ) -> Tuple: super().__init__(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = weight_decay_rate __SCREAMING_SNAKE_CASE = include_in_weight_decay __SCREAMING_SNAKE_CASE = exclude_from_weight_decay @classmethod def UpperCAmelCase_ ( cls : Optional[int] , UpperCAmelCase__ : Union[str, Any] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = {"WarmUp": WarmUp} return super(UpperCAmelCase__ , cls ).from_config(UpperCAmelCase__ , custom_objects=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] ) -> Dict: super(UpperCAmelCase__ , self )._prepare_local(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tf.constant( self.weight_decay_rate , name="adam_weight_decay_rate" ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] ) -> List[Any]: __SCREAMING_SNAKE_CASE = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]["weight_decay_rate"] , use_locking=self._use_locking , ) return tf.no_op() def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Any ) -> Optional[int]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = list(zip(*UpperCAmelCase__ ) ) return super(UpperCAmelCase__ , self ).apply_gradients(zip(UpperCAmelCase__ , UpperCAmelCase__ ) , name=UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any] ) -> Tuple: if apply_state is None: return self._decayed_lr_t[var_dtype], {} __SCREAMING_SNAKE_CASE = apply_state or {} __SCREAMING_SNAKE_CASE = apply_state.get((var_device, var_dtype) ) if coefficients is None: __SCREAMING_SNAKE_CASE = self._fallback_apply_state(UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[Any]=None ) -> Tuple: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self._get_lr(var.device , var.dtype.base_dtype , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self._decay_weights_op(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) with tf.control_dependencies([decay] ): return super(UpperCAmelCase__ , self )._resource_apply_dense(UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int]=None ) -> Dict: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self._get_lr(var.device , var.dtype.base_dtype , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self._decay_weights_op(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) with tf.control_dependencies([decay] ): return super(UpperCAmelCase__ , self )._resource_apply_sparse(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> Dict: __SCREAMING_SNAKE_CASE = super().get_config() config.update({"weight_decay_rate": self.weight_decay_rate} ) return config def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : List[Any] ) -> List[Any]: if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(UpperCAmelCase__ , UpperCAmelCase__ ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(UpperCAmelCase__ , UpperCAmelCase__ ) is not None: return False return True class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Optional[Any] ) -> str: __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = None @property def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: if self._accum_steps is None: __SCREAMING_SNAKE_CASE = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=UpperCAmelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def UpperCAmelCase_ ( self : Any ) -> Any: if not self._gradients: raise ValueError("The accumulator should be called first to initialize the gradients" ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self : Dict , UpperCAmelCase__ : Optional[int] ) -> Union[str, Any]: if not self._gradients: __SCREAMING_SNAKE_CASE = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(UpperCAmelCase__ ) , trainable=UpperCAmelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(UpperCAmelCase__ ) != len(self._gradients ): raise ValueError(F"""Expected {len(self._gradients )} gradients, but got {len(UpperCAmelCase__ )}""" ) for accum_gradient, gradient in zip(self._gradients , UpperCAmelCase__ ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(UpperCAmelCase__ ) self._accum_steps.assign_add(1 ) def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(UpperCAmelCase__ ) )
54
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
0
'''simple docstring''' from collections import defaultdict def __snake_case ( UpperCAmelCase_ : str , UpperCAmelCase_ : str ): lowerCamelCase_ = first_str.lower().strip() lowerCamelCase_ = second_str.lower().strip() # Remove whitespace lowerCamelCase_ = first_str.replace(" " , "" ) lowerCamelCase_ = second_str.replace(" " , "" ) # Strings of different lengths are not anagrams if len(UpperCAmelCase_ ) != len(UpperCAmelCase_ ): return False # Default values for count should be 0 lowerCamelCase_ = defaultdict(UpperCAmelCase_ ) # For each character in input strings, # increment count in the corresponding for i in range(len(UpperCAmelCase_ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() a_ : str = input("""Enter the first string """).strip() a_ : Dict = input("""Enter the second string """).strip() a_ : Dict = check_anagrams(input_a, input_b) print(f'''{input_a} and {input_b} are {'' if status else 'not '}anagrams.''')
55
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
0
'''simple docstring''' def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: # noqa: E741 '''simple docstring''' snake_case_ = len(__UpperCAmelCase ) snake_case_ = 0 snake_case_ = [0] * n snake_case_ = [False] * n snake_case_ = [False] * n def dfs(__UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ): if parent == root: out_edge_count += 1 snake_case_ = True snake_case_ = at for to in l[at]: if to == parent: pass elif not visited[to]: snake_case_ = dfs(__UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) snake_case_ = min(low[at], low[to] ) # AP found via bridge if at < low[to]: snake_case_ = True # AP found via cycle if at == low[to]: snake_case_ = True else: snake_case_ = min(low[at], __UpperCAmelCase ) return out_edge_count for i in range(__UpperCAmelCase ): if not visited[i]: snake_case_ = 0 snake_case_ = dfs(__UpperCAmelCase, __UpperCAmelCase, -1, __UpperCAmelCase ) snake_case_ = out_edge_count > 1 for x in range(len(__UpperCAmelCase ) ): if is_art[x] is True: print(__UpperCAmelCase ) # Adjacency list of graph a : Any = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
56
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase__ ( lowercase ): def __init__( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : str = dataset _UpperCamelCase : Optional[Any] = process _UpperCamelCase : Optional[Any] = params def __len__( self : Tuple ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.dataset[i] _UpperCamelCase : Dict = self.process(lowerCamelCase__ ,**self.params ) return processed class lowercase__ ( lowercase ): def __init__( self : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int]=None ): '''simple docstring''' _UpperCamelCase : Optional[int] = loader _UpperCamelCase : Tuple = infer _UpperCamelCase : List[str] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _UpperCamelCase : Any = None _UpperCamelCase : Union[str, Any] = loader_batch_size # Internal bookkeeping _UpperCamelCase : Optional[Any] = None _UpperCamelCase : str = None def __len__( self : List[str] ): '''simple docstring''' return len(self.loader ) def __iter__( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = iter(self.loader ) return self def UpperCamelCase_ ( self : Any ): '''simple docstring''' if isinstance(self._loader_batch_data ,torch.Tensor ): # Batch data is simple tensor, just fetch the slice _UpperCamelCase : Union[str, Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _UpperCamelCase : Union[str, Any] = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Convert ModelOutput to tuple first _UpperCamelCase : str = element.to_tuple() if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCamelCase__ ,lowerCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] ,torch.Tensor ): _UpperCamelCase : Dict = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] ,np.ndarray ): _UpperCamelCase : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] ,0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _UpperCamelCase : Optional[int] = None elif isinstance(element[self._loader_batch_index] ,torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] ,np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _UpperCamelCase : Optional[Any] = np.expand_dims(element[self._loader_batch_index] ,0 ) else: # This is typically a list, so no need to `unsqueeze`. _UpperCamelCase : Union[str, Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _UpperCamelCase : Optional[int] = self._loader_batch_data.__class__(lowerCamelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _UpperCamelCase : Tuple = next(self.iterator ) _UpperCamelCase : List[str] = self.infer(lowerCamelCase__ ,**self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : List[Any] = processed else: _UpperCamelCase : List[Any] = list(processed.keys() )[0] _UpperCamelCase : Optional[int] = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : int = len(lowerCamelCase__ ) else: _UpperCamelCase : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : int = observed_batch_size # Setting internal index to unwrap the batch _UpperCamelCase : Dict = processed _UpperCamelCase : str = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase__ ( lowercase ): def __init__( self : str ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Any=None ): '''simple docstring''' super().__init__(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def __iter__( self : Dict ): '''simple docstring''' _UpperCamelCase : str = iter(self.loader ) _UpperCamelCase : List[str] = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: _UpperCamelCase : Tuple = self.infer(next(self.iterator ) ,**self.params ) try: # Try to return next item _UpperCamelCase : Optional[Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) _UpperCamelCase : int = next(self.subiterator ) return processed class lowercase__ ( lowercase ): def __iter__( self : List[str] ): '''simple docstring''' _UpperCamelCase : Dict = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _UpperCamelCase : Dict = False _UpperCamelCase : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : Dict = self.loader_batch_item() _UpperCamelCase : List[str] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator while not is_last: _UpperCamelCase : List[Any] = self.infer(next(self.iterator ) ,**self.params ) if self.loader_batch_size is not None: if isinstance(lowerCamelCase__ ,torch.Tensor ): _UpperCamelCase : str = processed else: _UpperCamelCase : Any = list(processed.keys() )[0] _UpperCamelCase : Tuple = processed[key] if isinstance(lowerCamelCase__ ,lowerCamelCase__ ): _UpperCamelCase : Dict = len(lowerCamelCase__ ) else: _UpperCamelCase : Tuple = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _UpperCamelCase : Any = observed_batch_size _UpperCamelCase : List[Any] = processed _UpperCamelCase : int = 0 while self._loader_batch_index < self.loader_batch_size: _UpperCamelCase : List[Any] = self.loader_batch_item() _UpperCamelCase : Optional[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) if is_last: return accumulator else: _UpperCamelCase : Any = processed _UpperCamelCase : List[Any] = item.pop('is_last' ) accumulator.append(lowerCamelCase__ ) return accumulator class lowercase__ ( lowercase ): def __init__( self : Tuple ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : str = key def __len__( self : Dict ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Tuple ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class lowercase__ ( lowercase ): def __init__( self : List[Any] ,lowerCamelCase__ : Dataset ,lowerCamelCase__ : str ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = dataset _UpperCamelCase : Optional[Any] = keya _UpperCamelCase : str = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : List[str] ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
83
0
"""simple docstring""" import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right A : Union[str, Any] = 5_0_0_0_3 A : Any = 5_0_0_0_2 @require_sentencepiece @require_tokenizers class _UpperCamelCase ( lowerCAmelCase__ ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : str =PLBartTokenizer __UpperCAmelCase : Optional[Any] =None __UpperCAmelCase : Optional[Any] =False def snake_case ( self ): super().setUp() # We have a SentencePiece fixture for testing __lowerCAmelCase = PLBartTokenizer(__a , language_codes="base" , keep_accents=__a ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case ( self ): __lowerCAmelCase = PLBartTokenizer(__a , language_codes="base" , keep_accents=__a ) __lowerCAmelCase = tokenizer.tokenize("This is a test" ) self.assertListEqual(__a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__a ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __lowerCAmelCase = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __lowerCAmelCase = tokenizer.convert_tokens_to_ids(__a ) self.assertListEqual( __a , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __lowerCAmelCase = tokenizer.convert_ids_to_tokens(__a ) self.assertListEqual( __a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) __lowerCAmelCase = tokenizer.vocab_size __lowerCAmelCase = [tokenizer.convert_ids_to_tokens(__a ) for x in range(end - 4 , __a )] self.assertListEqual(__a , ["__java__", "__python__", "__en_XX__", "<mask>"] ) __lowerCAmelCase = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" __lowerCAmelCase = tokenizer(__a ).input_ids self.assertEqual( tokenizer.decode(__a , skip_special_tokens=__a , clean_up_tokenization_spaces=__a ) , __a , ) def snake_case ( self ): __lowerCAmelCase = PLBartTokenizer(__a , language_codes="multi" , keep_accents=__a ) __lowerCAmelCase = tokenizer.tokenize("This is a test" ) self.assertListEqual(__a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__a ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __lowerCAmelCase = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __lowerCAmelCase = tokenizer.convert_tokens_to_ids(__a ) self.assertListEqual( __a , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __lowerCAmelCase = tokenizer.convert_ids_to_tokens(__a ) self.assertListEqual( __a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) __lowerCAmelCase = tokenizer.vocab_size __lowerCAmelCase = [tokenizer.convert_ids_to_tokens(__a ) for x in range(end - 7 , __a )] self.assertListEqual( __a , ["__java__", "__python__", "__en_XX__", "__javascript__", "__php__", "__ruby__", "__go__"] ) __lowerCAmelCase = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" __lowerCAmelCase = tokenizer(__a ).input_ids self.assertEqual( tokenizer.decode(__a , skip_special_tokens=__a , clean_up_tokenization_spaces=__a ) , __a , ) @require_torch @require_sentencepiece @require_tokenizers class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Dict ="""uclanlp/plbart-python-en_XX""" __UpperCAmelCase : List[str] =[ """def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])""", """def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])""", ] __UpperCAmelCase : List[Any] =[ """Returns the maximum value of a b c.""", """Sums the values of a b c.""", ] __UpperCAmelCase : Optional[int] =[ 1_3_4, 5_4_5_2, 3_3_4_6_0, 3_3_4_4_1, 3_3_4_6_3, 3_3_4_6_5, 3_3_4_6_3, 3_3_4_4_9, 9_8_8, 2_0, 3_3_4_5_6, 1_9, 3_3_4_5_6, 7_7_1, 3_9, 4_2_5_8, 8_8_9, 3_3_1_8, 3_3_4_4_1, 3_3_4_6_3, 3_3_4_6_5, 3_3_4_6_3, 3_3_4_4_9, 2_4_7_1, 2, PYTHON_CODE, ] @classmethod def snake_case ( cls ): __lowerCAmelCase = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes="base" , src_lang="python" , tgt_lang="en_XX" ) __lowerCAmelCase = 1 return cls def snake_case ( self ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__java__"] , 5_00_01 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__python__"] , 5_00_02 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__en_XX__"] , 5_00_03 ) def snake_case ( self ): __lowerCAmelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __a ) def snake_case ( self ): self.assertIn(__a , self.tokenizer.all_special_ids ) __lowerCAmelCase = [EN_CODE, 90_37, 3_34_42, 57, 7_52, 1_53, 14, 56, 18, 9, 2] __lowerCAmelCase = self.tokenizer.decode(__a , skip_special_tokens=__a ) __lowerCAmelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__a ) self.assertEqual(__a , __a ) self.assertNotIn(self.tokenizer.eos_token , __a ) def snake_case ( self ): __lowerCAmelCase = ["def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])" * 20] self.assertIsInstance(src_text[0] , __a ) __lowerCAmelCase = 10 __lowerCAmelCase = self.tokenizer(__a , max_length=__a , truncation=__a ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __a ) self.assertEqual(len(__a ) , __a ) def snake_case ( self ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "__java__"] ) , [5_00_04, 5_00_01] ) def snake_case ( self ): __lowerCAmelCase = tempfile.mkdtemp() __lowerCAmelCase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__a ) __lowerCAmelCase = PLBartTokenizer.from_pretrained(__a ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __a ) @require_torch def snake_case ( self ): __lowerCAmelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__a , return_tensors="pt" ) __lowerCAmelCase = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , __a ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def snake_case ( self ): __lowerCAmelCase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__a , truncation=__a , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) __lowerCAmelCase = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(__a , __a ) self.assertEqual((2, 26) , batch.input_ids.shape ) self.assertEqual((2, 26) , batch.attention_mask.shape ) __lowerCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __a ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def snake_case ( self ): __lowerCAmelCase = self.tokenizer(self.src_text , padding=__a , truncation=__a , max_length=3 , return_tensors="pt" ) __lowerCAmelCase = self.tokenizer( text_target=self.tgt_text , padding=__a , truncation=__a , max_length=10 , return_tensors="pt" ) __lowerCAmelCase = targets["input_ids"] __lowerCAmelCase = shift_tokens_right(__a , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def snake_case ( self ): __lowerCAmelCase = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="java" ) self.assertEqual( nested_simplify(__a ) , { # A, test, EOS, en_XX "input_ids": [[1_50, 2_42, 2, 5_00_03]], "attention_mask": [[1, 1, 1, 1]], # java "forced_bos_token_id": 5_00_01, } , )
57
'''simple docstring''' import os from datetime import datetime as dt from github import Github snake_case_ : Any = [ 'good first issue', 'good second issue', 'good difficult issue', 'enhancement', 'new pipeline/model', 'new scheduler', 'wip', ] def A__ ( ): _UpperCamelCase : Tuple = Github(os.environ['GITHUB_TOKEN'] ) _UpperCamelCase : List[Any] = g.get_repo('huggingface/diffusers' ) _UpperCamelCase : List[Any] = repo.get_issues(state='open' ) for issue in open_issues: _UpperCamelCase : Dict = sorted(issue.get_comments() , key=lambda UpperCAmelCase_ : i.created_at , reverse=UpperCAmelCase_ ) _UpperCamelCase : List[str] = comments[0] if len(UpperCAmelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='closed' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='open' ) issue.remove_from_labels('stale' ) elif ( (dt.utcnow() - issue.updated_at).days > 2_3 and (dt.utcnow() - issue.created_at).days >= 3_0 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) issue.add_to_labels('stale' ) if __name__ == "__main__": main()
83
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class a_ ( metaclass=snake_case_ ): '''simple docstring''' UpperCamelCase = ['''torch''', '''torchsde'''] def __init__( self , *A , **A ) -> Union[str, Any]: requires_backends(self , ["""torch""", """torchsde"""] ) @classmethod def snake_case_( cls , *A , **A ) -> int: requires_backends(cls , ["""torch""", """torchsde"""] ) @classmethod def snake_case_( cls , *A , **A ) -> int: requires_backends(cls , ["""torch""", """torchsde"""] )
58
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase ) , """Tatoeba directory does not exist.""" ) class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Dict = self.resolver.write_model_card('opus-mt-he-en' ,dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
0
# DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class UpperCAmelCase ( A_ ): A__ : torch.FloatTensor A__ : Optional[torch.FloatTensor] = None def UpperCamelCase ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any=0.999 , __lowerCamelCase : Optional[Any]="cosine" , ): if alpha_transform_type == "cosine": def alpha_bar_fn(__lowerCamelCase : List[str] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__lowerCamelCase : List[str] ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) snake_case : Optional[int] = [] for i in range(__lowerCamelCase ): snake_case : Optional[int] = i / num_diffusion_timesteps snake_case : Tuple = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__lowerCamelCase ) / alpha_bar_fn(__lowerCamelCase ) , __lowerCamelCase ) ) return torch.tensor(__lowerCamelCase , dtype=torch.floataa ) class UpperCAmelCase ( A_ ,A_ ): A__ : Union[str, Any] = 1 @register_to_config def __init__(self : Optional[int] , snake_case__ : int = 10_00 , snake_case__ : float = 0.0001 , snake_case__ : float = 0.02 , snake_case__ : str = "linear" , snake_case__ : Optional[Union[np.ndarray, List[float]]] = None , snake_case__ : bool = True , snake_case__ : bool = True , snake_case__ : int = 0 , snake_case__ : str = "epsilon" , snake_case__ : float = 1.0 , **snake_case__ : Union[str, Any] , ) -> Optional[Any]: '''simple docstring''' if kwargs.get("set_alpha_to_one" , snake_case__ ) is not None: snake_case : Dict = ( "The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead." ) deprecate("set_alpha_to_one" , "1.0.0" , snake_case__ , standard_warn=snake_case__ ) snake_case : Optional[Any] = kwargs["set_alpha_to_one"] if trained_betas is not None: snake_case : Optional[Any] = torch.tensor(snake_case__ , dtype=torch.floataa ) elif beta_schedule == "linear": snake_case : List[str] = torch.linspace(snake_case__ , snake_case__ , snake_case__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case : Tuple = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , snake_case__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case : Tuple = betas_for_alpha_bar(snake_case__ ) else: raise NotImplementedError(f"""{beta_schedule} does is not implemented for {self.__class__}""" ) snake_case : Tuple = 1.0 - self.betas snake_case : int = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. snake_case : Any = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution snake_case : List[Any] = 1.0 # setable values snake_case : Optional[int] = None snake_case : List[Any] = torch.from_numpy(np.arange(0 , snake_case__ ).copy().astype(np.intaa ) ) def _SCREAMING_SNAKE_CASE (self : int , snake_case__ : torch.FloatTensor , snake_case__ : Optional[int] = None ) -> torch.FloatTensor: '''simple docstring''' return sample def _SCREAMING_SNAKE_CASE (self : List[str] , snake_case__ : int , snake_case__ : Union[str, torch.device] = None ) -> Any: '''simple docstring''' if num_inference_steps > self.config.num_train_timesteps: raise ValueError( f"""`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:""" f""" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle""" f""" maximal {self.config.num_train_timesteps} timesteps.""" ) snake_case : Union[str, Any] = num_inference_steps snake_case : Union[str, Any] = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case : Optional[Any] = (np.arange(0 , snake_case__ ) * step_ratio).round().copy().astype(np.intaa ) snake_case : Optional[Any] = torch.from_numpy(snake_case__ ).to(snake_case__ ) self.timesteps += self.config.steps_offset def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : torch.FloatTensor , snake_case__ : int , snake_case__ : torch.FloatTensor , snake_case__ : float = 0.0 , snake_case__ : bool = False , snake_case__ : Optional[torch.FloatTensor] = None , snake_case__ : bool = True , ) -> Union[DDIMSchedulerOutput, Tuple]: '''simple docstring''' snake_case : Tuple = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process snake_case : str = self.alphas_cumprod[timestep] snake_case : Union[str, Any] = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) snake_case : Optional[int] = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": snake_case : List[str] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 snake_case : Optional[int] = model_output elif self.config.prediction_type == "sample": snake_case : Optional[Any] = model_output snake_case : Optional[Any] = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": snake_case : Dict = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output snake_case : str = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or""" " `v_prediction`" ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: snake_case : Tuple = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf snake_case : Any = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf snake_case : str = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=snake_case__ , pred_original_sample=snake_case__ ) def __len__(self : Dict ) -> Optional[Any]: '''simple docstring''' return self.config.num_train_timesteps
59
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
0