code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
_UpperCamelCase = "\n# Installazione di Transformers\n! pip install transformers datasets\n# Per installare dalla fonte invece dell'ultima versione rilasciata, commenta il comando sopra e\n# rimuovi la modalità commento al comando seguente.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" _UpperCamelCase = [{"type": "code", "content": INSTALL_CONTENT}] _UpperCamelCase = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
275
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
1
from math import factorial class __lowercase : def __init__( self , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = real if isinstance(A_ , A_ ): __lowerCAmelCase : int = [1] * rank else: __lowerCAmelCase : Optional[Any] = rank def __repr__( self ) ->List[str]: '''simple docstring''' return ( f"""{self.real}+""" f"""{"+".join(str(A_ )+"E"+str(n+1 )for n,dual in enumerate(self.duals ) )}""" ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : str = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real , A_ ) def __add__( self , A_ ) ->str: '''simple docstring''' if not isinstance(A_ , A_ ): return Dual(self.real + other , self.duals ) __lowerCAmelCase : Any = self.duals.copy() __lowerCAmelCase : str = other.duals.copy() if len(A_ ) > len(A_ ): o_dual.extend([1] * (len(A_ ) - len(A_ )) ) elif len(A_ ) < len(A_ ): s_dual.extend([1] * (len(A_ ) - len(A_ )) ) __lowerCAmelCase : List[Any] = [] for i in range(len(A_ ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real , A_ ) _UpperCamelCase = __add__ def __sub__( self , A_ ) ->List[Any]: '''simple docstring''' return self + other * -1 def __mul__( self , A_ ) ->str: '''simple docstring''' if not isinstance(A_ , A_ ): __lowerCAmelCase : List[Any] = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other , A_ ) __lowerCAmelCase : List[str] = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real , A_ ) _UpperCamelCase = __mul__ def __truediv__( self , A_ ) ->Tuple: '''simple docstring''' if not isinstance(A_ , A_ ): __lowerCAmelCase : Tuple = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other , A_ ) raise ValueError def __floordiv__( self , A_ ) ->Union[str, Any]: '''simple docstring''' if not isinstance(A_ , A_ ): __lowerCAmelCase : Optional[int] = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other , A_ ) raise ValueError def __pow__( self , A_ ) ->Optional[int]: '''simple docstring''' if n < 0 or isinstance(A_ , A_ ): raise ValueError('''power must be a positive integer''' ) if n == 0: return 1 if n == 1: return self __lowerCAmelCase : Optional[Any] = self for _ in range(n - 1 ): x *= self return x def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): if not callable(lowercase__ ): raise ValueError('''differentiate() requires a function as input for func''' ) if not isinstance(lowercase__ , (float, int) ): raise ValueError('''differentiate() requires a float as input for position''' ) if not isinstance(lowercase__ , lowercase__ ): raise ValueError('''differentiate() requires an int as input for order''' ) __lowerCAmelCase : Tuple = Dual(lowercase__ , 1 ) __lowerCAmelCase : str = func(lowercase__ ) if order == 0: return result.real return result.duals[order - 1] * factorial(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod() def _lowercase ( lowercase__ ): return y**2 * y**4 print(differentiate(f, 9, 2))
275
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _UpperCamelCase = { "AI-Sweden/gpt-sw3-126m": 2048, "AI-Sweden/gpt-sw3-350m": 2048, "AI-Sweden/gpt-sw3-1.6b": 2048, "AI-Sweden/gpt-sw3-6.7b": 2048, "AI-Sweden/gpt-sw3-20b": 2048, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_ , A_=False , A_=False , A_=False , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCAmelCase : int = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCAmelCase : Union[str, Any] = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCAmelCase : str = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCAmelCase : Any = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token __lowerCAmelCase : int = eos_token if bos_token is None else bos_token else: __lowerCAmelCase : Optional[int] = '''<pad>''' if pad_token is None else pad_token __lowerCAmelCase : List[str] = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=A_ , remove_space=A_ , keep_accents=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Union[str, Any] = do_lower_case __lowerCAmelCase : Union[str, Any] = remove_space __lowerCAmelCase : int = keep_accents __lowerCAmelCase : Union[str, Any] = vocab_file __lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # Used for whitespace normalization in input texts # fmt : off __lowerCAmelCase : List[Any] = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCAmelCase : int = re.compile( f"""[{"".join(map(A_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.__dict__.copy() __lowerCAmelCase : List[Any] = None return state def __setstate__( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.non_printing_characters_re.sub('''''' , A_ ) # Normalize whitespaces __lowerCAmelCase : List[str] = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCAmelCase : Tuple = unicodedata.normalize('''NFC''' , A_ ) return text def UpperCamelCase__ ( self , A_ , **A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = self.preprocess_text(A_ ) return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' return self.sp_model.PieceToId(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.IdToPiece(A_ ) @staticmethod def UpperCamelCase__ ( A_ ) ->str: '''simple docstring''' return out_string def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = [] __lowerCAmelCase : Tuple = '''''' __lowerCAmelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string def UpperCamelCase__ ( self ) ->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : str = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Any = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = False ) ->Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[Any] = self.preprocess_text(A_ ) __lowerCAmelCase : Dict = self.sp_model.encode(A_ ) else: __lowerCAmelCase : Dict = [self.preprocess_text(A_ ) for t in text] __lowerCAmelCase : Optional[int] = self.sp_model.encode(A_ ) if return_tensors is True or return_tensors == "pt": __lowerCAmelCase : Tuple = torch.tensor(A_ ) return token_ids def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.decode(A_ ) def UpperCamelCase__ ( self , A_ ) ->List[int]: '''simple docstring''' __lowerCAmelCase : int = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCAmelCase : Any = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(A_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=A_ )
275
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _UpperCamelCase = None _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = "▁" _UpperCamelCase = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} _UpperCamelCase = { "vocab_file": {"google/pegasus-xsum": "https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"}, "tokenizer_file": { "google/pegasus-xsum": "https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json" }, } _UpperCamelCase = { "google/pegasus-xsum": 512, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = PegasusTokenizer _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_=None , A_=None , A_="<pad>" , A_="</s>" , A_="<unk>" , A_="<mask_2>" , A_="<mask_1>" , A_=None , A_=103 , **A_ , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = offset if additional_special_tokens is not None: if not isinstance(A_ , A_ ): raise TypeError( f"""additional_special_tokens should be of type {type(A_ )}, but is""" f""" {type(A_ )}""" ) __lowerCAmelCase : Union[str, Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f"""<unk_{i}>""" for i in range(len(A_ ) , self.offset - 1 ) ] if len(set(A_ ) ) != len(A_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) __lowerCAmelCase : Union[str, Any] = additional_special_tokens_extended else: __lowerCAmelCase : Any = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( A_ , tokenizer_file=A_ , pad_token=A_ , eos_token=A_ , unk_token=A_ , mask_token=A_ , mask_token_sent=A_ , offset=A_ , additional_special_tokens=A_ , **A_ , ) __lowerCAmelCase : List[str] = vocab_file __lowerCAmelCase : List[Any] = False if not self.vocab_file else True def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( '''There should be 3 special tokens: mask_token, pad_token, and eos_token +''' f""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def UpperCamelCase__ ( self , A_ , A_ = None , A_ = False ) ->List[int]: '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(A_ ) elif token_ids_a is None: return self._special_token_mask(A_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def UpperCamelCase__ ( self , A_ , A_=None ) ->List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Dict = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ): copyfile(self.vocab_file , A_ ) return (out_vocab_file,)
275
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """table-transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) ->Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowerCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(A_ , A_ ): __lowerCAmelCase : int = backbone_config.get('''model_type''' ) __lowerCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __lowerCAmelCase : Any = config_class.from_dict(A_ ) # set timm attributes to None __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = None, None, None __lowerCAmelCase : Tuple = use_timm_backbone __lowerCAmelCase : Optional[Any] = backbone_config __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = num_queries __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_ffn_dim __lowerCAmelCase : Optional[int] = encoder_layers __lowerCAmelCase : List[str] = encoder_attention_heads __lowerCAmelCase : str = decoder_ffn_dim __lowerCAmelCase : Union[str, Any] = decoder_layers __lowerCAmelCase : Any = decoder_attention_heads __lowerCAmelCase : Optional[int] = dropout __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : Tuple = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : List[str] = init_std __lowerCAmelCase : Tuple = init_xavier_std __lowerCAmelCase : Any = encoder_layerdrop __lowerCAmelCase : List[Any] = decoder_layerdrop __lowerCAmelCase : Optional[Any] = encoder_layers __lowerCAmelCase : Optional[Any] = auxiliary_loss __lowerCAmelCase : Optional[Any] = position_embedding_type __lowerCAmelCase : Tuple = backbone __lowerCAmelCase : Any = use_pretrained_backbone __lowerCAmelCase : int = dilation # Hungarian matcher __lowerCAmelCase : Dict = class_cost __lowerCAmelCase : List[str] = bbox_cost __lowerCAmelCase : int = giou_cost # Loss coefficients __lowerCAmelCase : Optional[Any] = mask_loss_coefficient __lowerCAmelCase : Tuple = dice_loss_coefficient __lowerCAmelCase : int = bbox_loss_coefficient __lowerCAmelCase : List[Any] = giou_loss_coefficient __lowerCAmelCase : int = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.d_model class __lowercase (_UpperCAmelCase ): _UpperCamelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def UpperCamelCase__ ( self ) ->float: '''simple docstring''' return 1e-5 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 12
275
1
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient _UpperCamelCase = WebClient(token=os.environ["CI_SLACK_BOT_TOKEN"]) def _lowercase ( lowercase__ ): __lowerCAmelCase : Any = test_results.split(''' ''' ) __lowerCAmelCase : List[Any] = 0 __lowerCAmelCase : int = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. __lowerCAmelCase : List[Any] = expressions[-2] if '''=''' in expressions[-1] else expressions[-1] for i, expression in enumerate(lowercase__ ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def _lowercase ( lowercase__ ): __lowerCAmelCase : str = {} __lowerCAmelCase : Dict = None __lowerCAmelCase : Any = False for line in failures_short_lines.split('''\n''' ): if re.search(r'''_ \[doctest\]''' , lowercase__ ): __lowerCAmelCase : int = True __lowerCAmelCase : Optional[Any] = line.split(''' ''' )[2] elif in_error and not line.split(''' ''' )[0].isdigit(): __lowerCAmelCase : Tuple = line __lowerCAmelCase : Optional[Any] = False return failures class __lowercase : def __init__( self , A_ , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = title __lowerCAmelCase : Optional[int] = doc_test_results['''time_spent'''].split(''',''' )[0] __lowerCAmelCase : Any = doc_test_results['''success'''] __lowerCAmelCase : Optional[int] = doc_test_results['''failures'''] __lowerCAmelCase : Union[str, Any] = self.n_success + self.n_failures # Failures and success of the modeling tests __lowerCAmelCase : Tuple = doc_test_results @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[int] = [self._time_spent] __lowerCAmelCase : Optional[Any] = 0 for time in time_spent: __lowerCAmelCase : int = time.split(''':''' ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(A_ ) == 1: __lowerCAmelCase : List[Any] = [0, 0, time_parts[0]] __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3600 + minutes * 60 + seconds __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = total_secs // 3600, (total_secs % 3600) // 60, total_secs % 60 return f"""{int(A_ )}h{int(A_ )}m{int(A_ )}s""" @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": f"""🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.""", "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": ( f"""There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in""" f""" {self.time}.""" ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : int = 40 __lowerCAmelCase : Optional[Any] = {k: v['''failed'''] for k, v in doc_test_results.items() if isinstance(A_ , A_ )} __lowerCAmelCase : Dict = '''''' for category, failures in category_failures.items(): if len(A_ ) == 0: continue if report != "": report += "\n\n" report += f"""*{category} failures*:""".ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(A_ ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f"""The following examples had failures:\n\n\n{report}\n""", }, } @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(A_ ) @staticmethod def UpperCamelCase__ ( ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = [ { '''type''': '''section''', '''text''': { '''type''': '''plain_text''', '''text''': '''There was an issue running the tests.''', }, '''accessory''': { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''Check Action results''', '''emoji''': True}, '''url''': f"""https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } ] print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(A_ )} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , text='''There was an issue running the tests.''' , blocks=A_ , ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(self.payload )} ) ) __lowerCAmelCase : List[Any] = f"""{self.n_failures} failures out of {self.n_tests} tests,""" if self.n_failures else '''All tests passed.''' __lowerCAmelCase : Optional[Any] = client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , blocks=self.payload , text=A_ , ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''''' for key, value in failures.items(): __lowerCAmelCase : Tuple = value[:200] + ''' [Truncated]''' if len(A_ ) > 250 else value failures_text += f"""*{key}*\n_{value}_\n\n""" __lowerCAmelCase : Union[str, Any] = job_name __lowerCAmelCase : Dict = {'''type''': '''section''', '''text''': {'''type''': '''mrkdwn''', '''text''': text}} if job_link is not None: __lowerCAmelCase : List[Any] = { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''GitHub Action job''', '''emoji''': True}, '''url''': job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' if self.thread_ts is None: raise ValueError('''Can only post reply if a post has been made.''' ) __lowerCAmelCase : Optional[Any] = self.doc_test_results.pop('''job_link''' ) self.doc_test_results.pop('''failures''' ) self.doc_test_results.pop('''success''' ) self.doc_test_results.pop('''time_spent''' ) __lowerCAmelCase : List[str] = sorted(self.doc_test_results.items() , key=lambda A_ : t[0] ) for job, job_result in sorted_dict: if len(job_result['''failures'''] ): __lowerCAmelCase : List[Any] = f"""*Num failures* :{len(job_result["failed"] )} \n""" __lowerCAmelCase : List[str] = job_result['''failures'''] __lowerCAmelCase : Any = self.get_reply_blocks(A_ , A_ , A_ , text=A_ ) print('''Sending the following reply''' ) print(json.dumps({'''blocks''': blocks} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , text=f"""Results for {job}""" , blocks=A_ , thread_ts=self.thread_ts['''ts'''] , ) time.sleep(1 ) def _lowercase ( ): __lowerCAmelCase : Dict = os.environ['''GITHUB_RUN_ID'''] __lowerCAmelCase : Optional[int] = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100""" __lowerCAmelCase : Tuple = requests.get(lowercase__ ).json() __lowerCAmelCase : List[Any] = {} try: jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) __lowerCAmelCase : List[str] = math.ceil((result['''total_count'''] - 1_0_0) / 1_0_0 ) for i in range(lowercase__ ): __lowerCAmelCase : Optional[int] = requests.get(url + f"""&page={i + 2}""" ).json() jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) return jobs except Exception as e: print('''Unknown error, could not fetch links.''' , lowercase__ ) return {} def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = {} if os.path.exists(lowercase__ ): __lowerCAmelCase : Optional[int] = os.listdir(lowercase__ ) for file in files: try: with open(os.path.join(lowercase__ , lowercase__ ) , encoding='''utf-8''' ) as f: __lowerCAmelCase : Optional[int] = f.read() except UnicodeDecodeError as e: raise ValueError(f"""Could not open {os.path.join(lowercase__ , lowercase__ )}.""" ) from e return _artifact def _lowercase ( ): class __lowercase : def __init__( self , A_ ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = name __lowerCAmelCase : Any = [] def __str__( self ) ->Optional[int]: '''simple docstring''' return self.name def UpperCamelCase__ ( self , A_ ) ->Any: '''simple docstring''' self.paths.append({'''name''': self.name, '''path''': path} ) __lowerCAmelCase : Dict[str, Artifact] = {} __lowerCAmelCase : Union[str, Any] = filter(os.path.isdir , os.listdir() ) for directory in directories: __lowerCAmelCase : List[Any] = directory if artifact_name not in _available_artifacts: __lowerCAmelCase : Optional[int] = Artifact(lowercase__ ) _available_artifacts[artifact_name].add_path(lowercase__ ) return _available_artifacts if __name__ == "__main__": _UpperCamelCase = get_job_links() _UpperCamelCase = retrieve_available_artifacts() _UpperCamelCase = collections.OrderedDict( [ ("*.py", "API Examples"), ("*.md", "MD Examples"), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' _UpperCamelCase = { v: { "failed": [], "failures": {}, } for v in docs.values() } # Link to the GitHub Action job _UpperCamelCase = github_actions_job_links.get("run_doctests") _UpperCamelCase = available_artifacts["doc_tests_gpu_test_reports"].paths[0] _UpperCamelCase = retrieve_artifact(artifact_path["name"]) if "stats" in artifact: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = handle_test_results(artifact["stats"]) _UpperCamelCase = failed _UpperCamelCase = success _UpperCamelCase = time_spent[1:-1] + ", " _UpperCamelCase = extract_first_line_failure(artifact["failures_short"]) for line in artifact["summary_short"].split("\n"): if re.search("FAILED", line): _UpperCamelCase = line.replace("FAILED ", "") _UpperCamelCase = line.split()[0].replace("\n", "") if "::" in line: _UpperCamelCase , _UpperCamelCase = line.split("::") else: _UpperCamelCase , _UpperCamelCase = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): _UpperCamelCase = docs[file_regex] doc_test_results[category]["failed"].append(test) _UpperCamelCase = all_failures[test] if test in all_failures else "N/A" _UpperCamelCase = failure break _UpperCamelCase = Message("🤗 Results of the doc tests.", doc_test_results) message.post() message.post_reply()
275
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = {} with open(lowercase__ , '''r''' ) as file: for line_number, line in enumerate(lowercase__ ): __lowerCAmelCase : Any = line.strip() if line: __lowerCAmelCase : Dict = line.split() __lowerCAmelCase : str = line_number __lowerCAmelCase : List[str] = words[0] __lowerCAmelCase : Any = value return result def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[Any] = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : Tuple = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : List[Any] = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : str = getattr(lowercase__ , lowercase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCAmelCase : Dict = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : str = shape_pointer.shape # let's reduce dimension __lowerCAmelCase : Any = value[0] else: __lowerCAmelCase : str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __lowerCAmelCase : List[str] = value elif weight_type == "weight_v": __lowerCAmelCase : int = value elif weight_type == "bias": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : int = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : Tuple = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCAmelCase : List[str] = '''.'''.join([key, hf_param_name] ) else: __lowerCAmelCase : Optional[int] = key __lowerCAmelCase : Union[str, Any] = value if '''lm_head''' in full_key else value[0] _UpperCamelCase = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None ): __lowerCAmelCase : Any = False for key, mapped_key in MAPPING.items(): __lowerCAmelCase : Tuple = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : List[str] = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : List[Any] = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : Any = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCAmelCase : int = '''weight''' else: __lowerCAmelCase : Any = None if hf_dict is not None: rename_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return is_used return is_used def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[Any] = fairseq_model.state_dict() __lowerCAmelCase : Tuple = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : int = True else: __lowerCAmelCase : Dict = load_wavaveca_layer(lowercase__ , lowercase__ , lowercase__ ) if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : List[str] = name.split('''.''' ) __lowerCAmelCase : Any = int(items[0] ) __lowerCAmelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=True , lowercase__=False ): if config_path is not None: __lowerCAmelCase : Union[str, Any] = WavaVecaConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Optional[int] = WavaVecaConfig() if is_seq_class: __lowerCAmelCase : Optional[Any] = read_txt_into_dict(lowercase__ ) __lowerCAmelCase : int = idalabel __lowerCAmelCase : Optional[int] = WavaVecaForSequenceClassification(lowercase__ ) __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) feature_extractor.save_pretrained(lowercase__ ) elif is_finetuned: if dict_path: __lowerCAmelCase : List[str] = Dictionary.load(lowercase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCAmelCase : List[Any] = target_dict.pad_index __lowerCAmelCase : List[Any] = target_dict.bos_index __lowerCAmelCase : Optional[int] = target_dict.eos_index __lowerCAmelCase : Any = len(target_dict.symbols ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''vocab.json''' ) if not os.path.isdir(lowercase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(lowercase__ ) ) return os.makedirs(lowercase__ , exist_ok=lowercase__ ) __lowerCAmelCase : Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : int = 1 with open(lowercase__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = WavaVecaCTCTokenizer( lowercase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=lowercase__ , ) __lowerCAmelCase : List[str] = True if config.feat_extract_norm == '''layer''' else False __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=lowercase__ , tokenizer=lowercase__ ) processor.save_pretrained(lowercase__ ) __lowerCAmelCase : str = WavaVecaForCTC(lowercase__ ) else: __lowerCAmelCase : Any = WavaVecaForPreTraining(lowercase__ ) if is_finetuned or is_seq_class: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCAmelCase : Union[str, Any] = argparse.Namespace(task='''audio_pretraining''' ) __lowerCAmelCase : str = fairseq.tasks.setup_task(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowercase__ ) __lowerCAmelCase : int = model[0].eval() recursively_load_weights(lowercase__ , lowercase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
275
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class __lowercase : _UpperCamelCase = PegasusConfig _UpperCamelCase = {} _UpperCamelCase = """gelu""" def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=False , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_=0.1 , A_=0.1 , A_=40 , A_=2 , A_=1 , A_=0 , ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : str = batch_size __lowerCAmelCase : Optional[Any] = seq_length __lowerCAmelCase : Optional[int] = is_training __lowerCAmelCase : Optional[int] = use_labels __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : Dict = num_hidden_layers __lowerCAmelCase : int = num_attention_heads __lowerCAmelCase : int = intermediate_size __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Union[str, Any] = attention_probs_dropout_prob __lowerCAmelCase : List[Any] = max_position_embeddings __lowerCAmelCase : List[str] = eos_token_id __lowerCAmelCase : Optional[int] = pad_token_id __lowerCAmelCase : Optional[Any] = bos_token_id def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : Any = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __lowerCAmelCase : List[Any] = prepare_pegasus_inputs_dict(A_ , A_ , A_ ) return config, inputs_dict def UpperCamelCase__ ( self , A_ , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = TFPegasusModel(config=A_ ).get_decoder() __lowerCAmelCase : Optional[Any] = inputs_dict['''input_ids'''] __lowerCAmelCase : Optional[Any] = input_ids[:1, :] __lowerCAmelCase : Union[str, Any] = inputs_dict['''attention_mask'''][:1, :] __lowerCAmelCase : Optional[int] = inputs_dict['''head_mask'''] __lowerCAmelCase : Optional[Any] = 1 # first forward pass __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , head_mask=A_ , use_cache=A_ ) __lowerCAmelCase, __lowerCAmelCase : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase : Dict = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase : Optional[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase : Dict = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase : Optional[int] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase : Any = model(A_ , attention_mask=A_ )[0] __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , past_key_values=A_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase : Dict = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase : str = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase : Union[str, Any] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A_ , A_ , rtol=1e-3 ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): if attention_mask is None: __lowerCAmelCase : int = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __lowerCAmelCase : Any = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __lowerCAmelCase : Any = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __lowerCAmelCase : Optional[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __lowerCAmelCase : List[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __lowercase (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () _UpperCamelCase = (TFPegasusForConditionalGeneration,) if is_tf_available() else () _UpperCamelCase = ( { """conversational""": TFPegasusForConditionalGeneration, """feature-extraction""": TFPegasusModel, """summarization""": TFPegasusForConditionalGeneration, """text2text-generation""": TFPegasusForConditionalGeneration, """translation""": TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = TFPegasusModelTester(self ) __lowerCAmelCase : str = ConfigTester(self , config_class=A_ ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A_ ) @require_sentencepiece @require_tokenizers @require_tf class __lowercase (unittest.TestCase ): _UpperCamelCase = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] _UpperCamelCase = [ """California's largest electricity provider has cut power to hundreds of thousands of customers in an effort to""" """ reduce the risk of wildfires.""", """N-Dubz have revealed they\'re \"grateful\" to have been nominated for four Mobo Awards.""", ] # differs slightly from pytorch, likely due to numerical differences in linear layers _UpperCamelCase = """google/pegasus-xsum""" @cached_property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[int] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def UpperCamelCase__ ( self , **A_ ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.translate_src_text(**A_ ) assert self.expected_text == generated_words def UpperCamelCase__ ( self , **A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.tokenizer(self.src_text , **A_ , padding=A_ , return_tensors='''tf''' ) __lowerCAmelCase : Dict = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=A_ , ) __lowerCAmelCase : Dict = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=A_ ) return generated_words @slow def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' self._assert_generated_batch_equal_expected()
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): def __init__( self , **A_ ) ->Optional[Any]: '''simple docstring''' requires_backends(self , ['''bs4'''] ) super().__init__(**A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = [] __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Optional[Any] = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag __lowerCAmelCase : Dict = parent.find_all(child.name , recursive=A_ ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(A_ ) else next(i for i, s in enumerate(A_ , 1 ) if s is child ) ) __lowerCAmelCase : Optional[Any] = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[str] = BeautifulSoup(A_ , '''html.parser''' ) __lowerCAmelCase : Tuple = [] __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Optional[Any] = [] for element in html_code.descendants: if type(A_ ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue __lowerCAmelCase : List[str] = html.unescape(A_ ).strip() if not text_in_this_tag: continue all_doc_strings.append(A_ ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.xpath_soup(A_ ) stringaxtag_seq.append(A_ ) stringaxsubs_seq.append(A_ ) if len(A_ ) != len(A_ ): raise ValueError('''Number of doc strings and xtags does not correspond''' ) if len(A_ ) != len(A_ ): raise ValueError('''Number of doc strings and xsubs does not correspond''' ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : Optional[Any] = '''''' for tagname, subs in zip(A_ , A_ ): xpath += f"""/{tagname}""" if subs != 0: xpath += f"""[{subs}]""" return xpath def __call__( self , A_ ) ->BatchFeature: '''simple docstring''' __lowerCAmelCase : int = False # Check that strings has a valid type if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[int] = True elif isinstance(A_ , (list, tuple) ): if len(A_ ) == 0 or isinstance(html_strings[0] , A_ ): __lowerCAmelCase : str = True if not valid_strings: raise ValueError( '''HTML strings must of type `str`, `List[str]` (batch of examples), ''' f"""but is of type {type(A_ )}.""" ) __lowerCAmelCase : List[str] = bool(isinstance(A_ , (list, tuple) ) and (isinstance(html_strings[0] , A_ )) ) if not is_batched: __lowerCAmelCase : List[str] = [html_strings] # Get nodes + xpaths __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[int] = [] for html_string in html_strings: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Tuple = self.get_three_from_single(A_ ) nodes.append(A_ ) __lowerCAmelCase : List[str] = [] for node, tag_list, sub_list in zip(A_ , A_ , A_ ): __lowerCAmelCase : str = self.construct_xpath(A_ , A_ ) xpath_strings.append(A_ ) xpaths.append(A_ ) # return as Dict __lowerCAmelCase : str = {'''nodes''': nodes, '''xpaths''': xpaths} __lowerCAmelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
275
def _lowercase ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __lowerCAmelCase : int = str(bin(lowercase__ ) )[2:] # remove the leading "0b" __lowerCAmelCase : Any = str(bin(lowercase__ ) )[2:] __lowerCAmelCase : List[str] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('''1''' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
275
1
import math def _lowercase ( lowercase__ , lowercase__ ): return math.pow(lowercase__ , 2 ) - a def _lowercase ( lowercase__ ): return 2 * x def _lowercase ( lowercase__ ): __lowerCAmelCase : str = 2.0 while start <= a: __lowerCAmelCase : Dict = math.pow(lowercase__ , 2 ) return start def _lowercase ( lowercase__ , lowercase__ = 9_9_9_9 , lowercase__ = 0.0_0_0_0_0_0_0_0_0_0_0_0_0_1 ): if a < 0: raise ValueError('''math domain error''' ) __lowerCAmelCase : List[Any] = get_initial_point(lowercase__ ) for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = value __lowerCAmelCase : Any = value - fx(lowercase__ , lowercase__ ) / fx_derivative(lowercase__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
275
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
1
def _lowercase ( lowercase__ = 6_0_0_8_5_1_4_7_5_1_4_3 ): try: __lowerCAmelCase : Dict = int(lowercase__ ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) __lowerCAmelCase : Tuple = 2 __lowerCAmelCase : str = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 __lowerCAmelCase : Optional[Any] = i while n % i == 0: __lowerCAmelCase : Optional[int] = n // i i += 1 return int(lowercase__ ) if __name__ == "__main__": print(F"{solution() = }")
275
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCAmelCase : Dict = Vector() def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(A_ ) , '''(0,0,0,0,0,1)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(A_ ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Vector([1, 2] ) __lowerCAmelCase : Optional[int] = Vector([1, 2, 3, 4, 5] ) __lowerCAmelCase : Optional[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCAmelCase : str = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : List[Any] = Vector([2, -1, 4] ) # for test of dot product __lowerCAmelCase : Optional[int] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : Any = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , A_ , A_ ) ) , '''(3,4,7)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) __lowerCAmelCase : Optional[Any] = x.copy() self.assertEqual(str(A_ ) , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(A_ ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Dict = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
275
1
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants _UpperCamelCase = Mapping[str, np.ndarray] _UpperCamelCase = Mapping[str, Any] # Is a nested dict. _UpperCamelCase = 0.01 @dataclasses.dataclass(frozen=_UpperCAmelCase ) class __lowercase : _UpperCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _UpperCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _UpperCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _UpperCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _UpperCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _UpperCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _UpperCamelCase = None # Templates used to generate this protein (prediction-only) _UpperCamelCase = None # Chain corresponding to each parent _UpperCamelCase = None def _lowercase ( lowercase__ ): __lowerCAmelCase : Optional[int] = r'''(\[[A-Z]+\]\n)''' __lowerCAmelCase : List[str] = [tag.strip() for tag in re.split(lowercase__ , lowercase__ ) if len(lowercase__ ) > 0] __lowerCAmelCase : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split('''\n''' ) for l in tags[1::2]] ) __lowerCAmelCase : List[str] = ["N", "CA", "C"] __lowerCAmelCase : Dict = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : int = None for g in groups: if "[PRIMARY]" == g[0]: __lowerCAmelCase : Optional[Any] = g[1][0].strip() for i in range(len(lowercase__ ) ): if seq[i] not in residue_constants.restypes: __lowerCAmelCase : Any = '''X''' # FIXME: strings are immutable __lowerCAmelCase : str = np.array( [residue_constants.restype_order.get(lowercase__ , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: __lowerCAmelCase : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(lowercase__ , g[1][axis].split() ) ) ) __lowerCAmelCase : List[str] = np.array(lowercase__ ) __lowerCAmelCase : str = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(lowercase__ ): __lowerCAmelCase : str = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: __lowerCAmelCase : Optional[int] = np.array(list(map({'''-''': 0, '''+''': 1}.get , g[1][0].strip() ) ) ) __lowerCAmelCase : Union[str, Any] = np.zeros( ( len(lowercase__ ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(lowercase__ ): __lowerCAmelCase : List[Any] = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=lowercase__ , atom_mask=lowercase__ , aatype=lowercase__ , residue_index=np.arange(len(lowercase__ ) ) , b_factors=lowercase__ , ) def _lowercase ( lowercase__ , lowercase__ = 0 ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Tuple = prot.remark if remark is not None: pdb_headers.append(f"""REMARK {remark}""" ) __lowerCAmelCase : Optional[Any] = prot.parents __lowerCAmelCase : Tuple = prot.parents_chain_index if parents is not None and parents_chain_index is not None: __lowerCAmelCase : List[str] = [p for i, p in zip(lowercase__ , lowercase__ ) if i == chain_id] if parents is None or len(lowercase__ ) == 0: __lowerCAmelCase : Dict = ['''N/A'''] pdb_headers.append(f"""PARENT {" ".join(lowercase__ )}""" ) return pdb_headers def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : str = pdb_str.split('''\n''' ) __lowerCAmelCase : Union[str, Any] = prot.remark if remark is not None: out_pdb_lines.append(f"""REMARK {remark}""" ) __lowerCAmelCase : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: __lowerCAmelCase : Tuple = [] if prot.parents_chain_index is not None: __lowerCAmelCase : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(lowercase__ ) , [] ) parent_dict[str(lowercase__ )].append(lowercase__ ) __lowerCAmelCase : int = max([int(lowercase__ ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): __lowerCAmelCase : Dict = parent_dict.get(str(lowercase__ ) , ['''N/A'''] ) parents_per_chain.append(lowercase__ ) else: parents_per_chain.append(list(prot.parents ) ) else: __lowerCAmelCase : Union[str, Any] = [['''N/A''']] def make_parent_line(lowercase__ ) -> str: return f"""PARENT {" ".join(lowercase__ )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) __lowerCAmelCase : Optional[Any] = 0 for i, l in enumerate(lowercase__ ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(lowercase__ ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(lowercase__ ): __lowerCAmelCase : Tuple = parents_per_chain[chain_counter] else: __lowerCAmelCase : Tuple = ['''N/A'''] out_pdb_lines.append(make_parent_line(lowercase__ ) ) return "\n".join(lowercase__ ) def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = residue_constants.restypes + ['''X'''] def res_atoa(lowercase__ ) -> str: return residue_constants.restype_atoa.get(restypes[r] , '''UNK''' ) __lowerCAmelCase : Tuple = residue_constants.atom_types __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Any = prot.atom_mask __lowerCAmelCase : Dict = prot.aatype __lowerCAmelCase : Any = prot.atom_positions __lowerCAmelCase : List[Any] = prot.residue_index.astype(np.intaa ) __lowerCAmelCase : int = prot.b_factors __lowerCAmelCase : Any = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError('''Invalid aatypes.''' ) __lowerCAmelCase : Optional[Any] = get_pdb_headers(lowercase__ ) if len(lowercase__ ) > 0: pdb_lines.extend(lowercase__ ) __lowerCAmelCase : Optional[int] = aatype.shape[0] __lowerCAmelCase : Any = 1 __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Union[str, Any] = string.ascii_uppercase __lowerCAmelCase : List[str] = None # Add all atom sites. for i in range(lowercase__ ): __lowerCAmelCase : Any = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(lowercase__ , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue __lowerCAmelCase : Optional[int] = '''ATOM''' __lowerCAmelCase : List[str] = atom_name if len(lowercase__ ) == 4 else f""" {atom_name}""" __lowerCAmelCase : List[Any] = '''''' __lowerCAmelCase : Optional[Any] = '''''' __lowerCAmelCase : Optional[int] = 1.0_0 __lowerCAmelCase : List[Any] = atom_name[0] # Protein supports only C, N, O, S, this works. __lowerCAmelCase : List[Any] = '''''' __lowerCAmelCase : Tuple = '''A''' if chain_index is not None: __lowerCAmelCase : int = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! __lowerCAmelCase : Union[str, Any] = ( f"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" f"""{res_name_a:>3} {chain_tag:>1}""" f"""{residue_index[i]:>4}{insertion_code:>1} """ f"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" f"""{occupancy:>6.2f}{b_factor:>6.2f} """ f"""{element:>2}{charge:>2}""" ) pdb_lines.append(lowercase__ ) atom_index += 1 __lowerCAmelCase : str = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: __lowerCAmelCase : Union[str, Any] = True __lowerCAmelCase : Optional[Any] = chain_index[i + 1] if should_terminate: # Close the chain. __lowerCAmelCase : Tuple = '''TER''' __lowerCAmelCase : Optional[int] = ( f"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(lowercase__ ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(lowercase__ , lowercase__ ) ) pdb_lines.append('''END''' ) pdb_lines.append('''''' ) return "\n".join(lowercase__ ) def _lowercase ( lowercase__ ): return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def _lowercase ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = None , lowercase__ = None , lowercase__ = None , ): return Protein( aatype=features['''aatype'''] , atom_positions=result['''final_atom_positions'''] , atom_mask=result['''final_atom_mask'''] , residue_index=features['''residue_index'''] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result['''final_atom_mask'''] ) , chain_index=lowercase__ , remark=lowercase__ , parents=lowercase__ , parents_chain_index=lowercase__ , )
275
def _lowercase ( lowercase__ , lowercase__ ): if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
275
1
def _lowercase ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
275
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
1
import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : Dict = emb.weight.shape __lowerCAmelCase : int = nn.Linear(lowercase__ , lowercase__ , bias=lowercase__ ) __lowerCAmelCase : Tuple = emb.weight.data return lin_layer def _lowercase ( lowercase__ , lowercase__=None ): __lowerCAmelCase : Union[str, Any] = {} for old_key in state_dict.keys(): __lowerCAmelCase : Any = old_key if "moe_layer.experts." in key: if expert_idx is not None: __lowerCAmelCase : int = key.replace('''moe_layer.experts.0''' , f"""ffn.experts.expert_{expert_idx}""" ) else: __lowerCAmelCase : Tuple = key.replace('''moe_layer.experts.''' , '''ffn.experts.expert_''' ) if "gate" in key: __lowerCAmelCase : List[Any] = key.replace('''.moe_layer.gate.wg''' , '''.ffn.router.classifier''' ) if "fc2" and "experts" not in key: __lowerCAmelCase : int = key.replace('''.fc2.''' , '''.ffn.fc2.''' ) if "fc1" and "experts" not in key: __lowerCAmelCase : Tuple = key.replace('''.fc1.''' , '''.ffn.fc1.''' ) if ".encoder_attn." in key: __lowerCAmelCase : Tuple = key.replace('''.encoder_attn.''' , '''.cross_attention.''' ) if "encoder_attn_layer_norm" in key: __lowerCAmelCase : Optional[int] = key.replace('''encoder_attn_layer_norm''' , '''cross_attention_layer_norm''' ) if "final_layer_norm" in key: __lowerCAmelCase : Optional[Any] = key.replace('''final_layer_norm''' , '''ff_layer_norm''' ) __lowerCAmelCase : Optional[Any] = state_dict[old_key] return new_dict def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = WEIGHTS_NAME ): __lowerCAmelCase : Tuple = [] __lowerCAmelCase : Optional[Any] = 0 os.makedirs(lowercase__ , exist_ok=lowercase__ ) for expert in range(lowercase__ ): __lowerCAmelCase : Optional[Any] = switch_checkpoint_path + f"""-rank-{expert}.pt""" if os.path.isfile(lowercase__ ): __lowerCAmelCase : List[str] = torch.load(lowercase__ )['''model'''] remove_ignore_keys_(lowercase__ ) __lowerCAmelCase : int = rename_fairseq_keys(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = os.path.join( lowercase__ , weights_name.replace('''.bin''' , f"""-{len(lowercase__ )+1:05d}-of-???.bin""" ) ) torch.save(lowercase__ , lowercase__ ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(lowercase__ )[0]].dtype ) # Add the last block __lowerCAmelCase : List[Any] = os.path.join(lowercase__ , weights_name.replace('''.bin''' , f"""-{len(lowercase__ )+1:05d}-of-???.bin""" ) ) __lowerCAmelCase : Optional[int] = torch.load(switch_checkpoint_path + '''-shared.pt''' )['''model'''] remove_ignore_keys_(lowercase__ ) __lowerCAmelCase : Optional[int] = rename_fairseq_keys(lowercase__ , lowercase__ ) __lowerCAmelCase : int = shared_weights['''decoder.embed_tokens.weight'''] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(lowercase__ ) == 1: __lowerCAmelCase : str = os.path.join(lowercase__ , lowercase__ ) torch.save(lowercase__ , lowercase__ ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(lowercase__ , lowercase__ ) # Otherwise, let's build the index __lowerCAmelCase : str = {} for idx, shard in enumerate(lowercase__ ): __lowerCAmelCase : Any = weights_name.replace('''.bin''' , f"""-{idx+1:05d}-of-{len(lowercase__ ):05d}.bin""" ) __lowerCAmelCase : str = os.path.join(lowercase__ , weights_name.replace('''.bin''' , f"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(lowercase__ , os.path.join(lowercase__ , lowercase__ ) ) for key in shard: __lowerCAmelCase : Optional[int] = shard_file # Add the metadata __lowerCAmelCase : List[Any] = {'''total_size''': total_size} __lowerCAmelCase : Union[str, Any] = {'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(lowercase__ , lowercase__ ) , '''w''' , encoding='''utf-8''' ) as f: __lowerCAmelCase : List[Any] = json.dumps(lowercase__ , indent=2 , sort_keys=lowercase__ ) + '''\n''' f.write(lowercase__ ) return metadata, index if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--nllb_moe_checkpoint_path", default="/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000", type=str, required=False, help="Path to a directory containing a folder per layer. Follows the original Google format.", ) parser.add_argument("--dtype", default="float32", type=str, required=False, help="dtype of the saved model") parser.add_argument( "--pytorch_dump_folder_path", default="/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b", type=str, required=False, help="Path to the output pytorch model.", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase , _UpperCamelCase = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 128, args.dtype, ) _UpperCamelCase = NllbMoeConfig.from_pretrained( "facebook/nllb-200-3.3B", encoder_sparse_step=4, decoder_sparse_step=4, num_experts=128 ) config.save_pretrained(args.pytorch_dump_folder_path) _UpperCamelCase = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print("Done") model.save_pretrained(args.pytorch_dump_folder_path)
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1
# Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union _UpperCamelCase = re.compile(r"^(?P<major>\d+)" r"\.(?P<minor>\d+)" r"\.(?P<patch>\d+)$") @total_ordering @dataclass class __lowercase : _UpperCamelCase = 42 _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = _str_to_version_tuple(self.version_str ) def __repr__( self ) ->Optional[int]: '''simple docstring''' return f"""{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}""" @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' return self.major, self.minor, self.patch def UpperCamelCase__ ( self , A_ ) ->Any: '''simple docstring''' if isinstance(A_ , A_ ): return Version(A_ ) elif isinstance(A_ , A_ ): return other raise TypeError(f"""{other} (type {type(A_ )}) cannot be compared to version.""" ) def __eq__( self , A_ ) ->Optional[Any]: '''simple docstring''' try: __lowerCAmelCase : Optional[int] = self._validate_operand(A_ ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self , A_ ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self._validate_operand(A_ ) return self.tuple < other.tuple def __hash__( self ) ->List[Any]: '''simple docstring''' return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def UpperCamelCase__ ( cls , A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Tuple = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.version_str def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = _VERSION_REG.match(lowercase__ ) if not res: raise ValueError(f"""Invalid version '{version_str}'. Format should be x.y.z with {{x,y,z}} being digits.""" ) return tuple(int(lowercase__ ) for v in [res.group('''major''' ), res.group('''minor''' ), res.group('''patch''' )] ) def _lowercase ( lowercase__ ): return ".".join(str(lowercase__ ) for v in version_tuple )
275
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ConsistencyModelPipeline _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt _UpperCamelCase = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[Any] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def UpperCamelCase__ ( self , A_=False ) ->Dict: '''simple docstring''' if class_cond: __lowerCAmelCase : List[str] = self.dummy_cond_unet else: __lowerCAmelCase : Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->Tuple: '''simple docstring''' if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : str = torch.manual_seed(A_ ) else: __lowerCAmelCase : Dict = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Tuple = self.get_dummy_components() __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : str = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : str = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : int = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Any = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : Union[str, Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : List[str] = 1 __lowerCAmelCase : Dict = None __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , A_=0 , A_=False , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = torch.manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __lowerCAmelCase : List[str] = self.get_fixed_latents(seed=A_ , device=A_ , dtype=A_ , shape=A_ ) __lowerCAmelCase : Union[str, Any] = latents return inputs def UpperCamelCase__ ( self , A_=0 , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->Optional[int]: '''simple docstring''' if type(A_ ) == str: __lowerCAmelCase : int = torch.device(A_ ) __lowerCAmelCase : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Union[str, Any] = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) return latents def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = self.get_inputs() __lowerCAmelCase : Any = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.0_888, 0.0_881, 0.0_666, 0.0_479, 0.0_292, 0.0_195, 0.0_201, 0.0_163, 0.0_254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = self.get_inputs() __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : str = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.0_340, 0.0_152, 0.0_063, 0.0_267, 0.0_221, 0.0_107, 0.0_416, 0.0_186, 0.0_217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_inputs(get_fixed_latents=A_ , device=A_ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.1_875, 0.1_428, 0.1_289, 0.2_151, 0.2_092, 0.1_477, 0.1_877, 0.1_641, 0.1_353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Union[str, Any] = self.get_inputs(get_fixed_latents=A_ , device=A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : int = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : str = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.1_663, 0.1_948, 0.2_275, 0.1_680, 0.1_204, 0.1_245, 0.1_858, 0.1_338, 0.2_095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
275
1
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets _UpperCamelCase = "\\n@inproceedings{snover-etal-2006-study,\n title = \"A Study of Translation Edit Rate with Targeted Human Annotation\",\n author = \"Snover, Matthew and\n Dorr, Bonnie and\n Schwartz, Rich and\n Micciulla, Linnea and\n Makhoul, John\",\n booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\",\n month = aug # \" 8-12\",\n year = \"2006\",\n address = \"Cambridge, Massachusetts, USA\",\n publisher = \"Association for Machine Translation in the Americas\",\n url = \"https://aclanthology.org/2006.amta-papers.25\",\n pages = \"223--231\",\n}\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" _UpperCamelCase = "\\nTER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a\nhypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu\n(https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found\nhere: https://github.com/jhclark/tercom.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information.\n" _UpperCamelCase = "\nProduces TER scores alongside the number of edits and reference length.\n\nArgs:\n predictions (list of str): The system stream (a sequence of segments).\n references (list of list of str): A list of one or more reference streams (each a sequence of segments).\n normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters,\n as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana.\n Only applies if `normalized = True`. Defaults to `False`.\n case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`.\n\nReturns:\n 'score' (float): TER score (num_edits / sum_ref_lengths * 100)\n 'num_edits' (int): The cumulative number of edits\n 'ref_length' (float): The cumulative average reference length\n\nExamples:\n Example 1:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\",\n ... \"What did the TER metric user say to the developer?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"],\n ... [\"Your jokes are...\", \"...TERrible\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0}\n\n Example 2:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0}\n\n Example 3:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... normalized=True,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5}\n\n Example 4:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0}\n\n Example 5:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\",\n ... \"What did the TER metric user say to the developer?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"],\n ... [\"Your jokes are...\", \"...TERrible\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase (datasets.Metric ): def UpperCamelCase__ ( self ) ->str: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse('''1.4.12''' ): raise ImportWarning( '''To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n''' '''You can install it with `pip install "sacrebleu>=1.4.12"`.''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''http://www.cs.umd.edu/~snover/tercom/''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=['''https://github.com/mjpost/sacreBLEU#ter'''] , reference_urls=[ '''https://github.com/jhclark/tercom''', ] , ) def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = False , A_ = False , A_ = False , ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = len(references[0] ) if any(len(A_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) __lowerCAmelCase : List[str] = [[refs[i] for refs in references] for i in range(A_ )] __lowerCAmelCase : Optional[int] = TER( normalized=A_ , no_punct=A_ , asian_support=A_ , case_sensitive=A_ , ) __lowerCAmelCase : Union[str, Any] = sb_ter.corpus_score(A_ , A_ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
275
from collections import deque from .hash_table import HashTable class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->int: '''simple docstring''' super().__init__(*A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(A_ ) __lowerCAmelCase : int = self.values[key] def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return ( sum(self.charge_factor - len(A_ ) for slot in self.values ) / self.size_table * self.charge_factor ) def UpperCamelCase__ ( self , A_ , A_=None ) ->str: '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(A_ ) == 0 ): return key return super()._collision_resolution(A_ , A_ )
275
1
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) set_seed(770) _UpperCamelCase = { "c_attn": "att_proj", "c_proj": "out_proj", "c_fc": "in_proj", "transformer.": "", "h.": "layers.", "ln_1": "layernorm_1", "ln_2": "layernorm_2", "ln_f": "layernorm_final", "wpe": "position_embeds_layer", "wte": "input_embeds_layer", } _UpperCamelCase = { "text_small": { "repo_id": "suno/bark", "file_name": "text.pt", }, "coarse_small": { "repo_id": "suno/bark", "file_name": "coarse.pt", }, "fine_small": { "repo_id": "suno/bark", "file_name": "fine.pt", }, "text": { "repo_id": "suno/bark", "file_name": "text_2.pt", }, "coarse": { "repo_id": "suno/bark", "file_name": "coarse_2.pt", }, "fine": { "repo_id": "suno/bark", "file_name": "fine_2.pt", }, } _UpperCamelCase = os.path.dirname(os.path.abspath(__file__)) _UpperCamelCase = os.path.join(os.path.expanduser("~"), ".cache") _UpperCamelCase = os.path.join(os.getenv("XDG_CACHE_HOME", default_cache_dir), "suno", "bark_v0") def _lowercase ( lowercase__ , lowercase__=False ): __lowerCAmelCase : Dict = model_type if use_small: key += "_small" return os.path.join(lowercase__ , REMOTE_MODEL_PATHS[key]['''file_name'''] ) def _lowercase ( lowercase__ , lowercase__ ): os.makedirs(lowercase__ , exist_ok=lowercase__ ) hf_hub_download(repo_id=lowercase__ , filename=lowercase__ , local_dir=lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=False , lowercase__="text" ): if model_type == "text": __lowerCAmelCase : str = BarkSemanticModel __lowerCAmelCase : Union[str, Any] = BarkSemanticConfig __lowerCAmelCase : int = BarkSemanticGenerationConfig elif model_type == "coarse": __lowerCAmelCase : str = BarkCoarseModel __lowerCAmelCase : List[Any] = BarkCoarseConfig __lowerCAmelCase : List[Any] = BarkCoarseGenerationConfig elif model_type == "fine": __lowerCAmelCase : Union[str, Any] = BarkFineModel __lowerCAmelCase : Tuple = BarkFineConfig __lowerCAmelCase : Any = BarkFineGenerationConfig else: raise NotImplementedError() __lowerCAmelCase : Tuple = f"""{model_type}_small""" if use_small else model_type __lowerCAmelCase : Optional[int] = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(lowercase__ ): logger.info(f"""{model_type} model not found, downloading into `{CACHE_DIR}`.""" ) _download(model_info['''repo_id'''] , model_info['''file_name'''] ) __lowerCAmelCase : Tuple = torch.load(lowercase__ , map_location=lowercase__ ) # this is a hack __lowerCAmelCase : Optional[Any] = checkpoint['''model_args'''] if "input_vocab_size" not in model_args: __lowerCAmelCase : Optional[int] = model_args['''vocab_size'''] __lowerCAmelCase : Dict = model_args['''vocab_size'''] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments __lowerCAmelCase : int = model_args.pop('''n_head''' ) __lowerCAmelCase : str = model_args.pop('''n_embd''' ) __lowerCAmelCase : Any = model_args.pop('''n_layer''' ) __lowerCAmelCase : Dict = ConfigClass(**checkpoint['''model_args'''] ) __lowerCAmelCase : str = ModelClass(config=lowercase__ ) __lowerCAmelCase : Optional[int] = GenerationConfigClass() __lowerCAmelCase : List[Any] = model_generation_config __lowerCAmelCase : List[str] = checkpoint['''model'''] # fixup checkpoint __lowerCAmelCase : Any = '''_orig_mod.''' for k, v in list(state_dict.items() ): if k.startswith(lowercase__ ): # replace part of the key with corresponding layer name in HF implementation __lowerCAmelCase : Any = k[len(lowercase__ ) :] for old_layer_name in new_layer_name_dict: __lowerCAmelCase : Dict = new_k.replace(lowercase__ , new_layer_name_dict[old_layer_name] ) __lowerCAmelCase : str = state_dict.pop(lowercase__ ) __lowerCAmelCase : str = set(state_dict.keys() ) - set(model.state_dict().keys() ) __lowerCAmelCase : Any = {k for k in extra_keys if not k.endswith('''.attn.bias''' )} __lowerCAmelCase : List[Any] = set(model.state_dict().keys() ) - set(state_dict.keys() ) __lowerCAmelCase : Union[str, Any] = {k for k in missing_keys if not k.endswith('''.attn.bias''' )} if len(lowercase__ ) != 0: raise ValueError(f"""extra keys found: {extra_keys}""" ) if len(lowercase__ ) != 0: raise ValueError(f"""missing keys: {missing_keys}""" ) model.load_state_dict(lowercase__ , strict=lowercase__ ) __lowerCAmelCase : Tuple = model.num_parameters(exclude_embeddings=lowercase__ ) __lowerCAmelCase : Optional[Any] = checkpoint['''best_val_loss'''].item() logger.info(f"""model loaded: {round(n_params/1E6 , 1 )}M params, {round(lowercase__ , 3 )} loss""" ) model.eval() model.to(lowercase__ ) del checkpoint, state_dict return model def _lowercase ( lowercase__ , lowercase__=False , lowercase__="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() __lowerCAmelCase : Any = '''cpu''' # do conversion on cpu __lowerCAmelCase : int = _get_ckpt_path(lowercase__ , use_small=lowercase__ ) __lowerCAmelCase : List[str] = _load_model(lowercase__ , lowercase__ , model_type=lowercase__ , use_small=lowercase__ ) # load bark initial model __lowerCAmelCase : str = _bark_load_model(lowercase__ , '''cpu''' , model_type=lowercase__ , use_small=lowercase__ ) if model_type == "text": __lowerCAmelCase : Optional[Any] = bark_model['''model'''] if model.num_parameters(exclude_embeddings=lowercase__ ) != bark_model.get_num_params(): raise ValueError('''initial and new models don\'t have the same number of parameters''' ) # check if same output as the bark model __lowerCAmelCase : List[Any] = 5 __lowerCAmelCase : Tuple = 1_0 if model_type in ["text", "coarse"]: __lowerCAmelCase : Dict = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) __lowerCAmelCase : Any = bark_model(lowercase__ )[0] __lowerCAmelCase : Tuple = model(lowercase__ ) # take last logits __lowerCAmelCase : Optional[int] = output_new_model_total.logits[:, [-1], :] else: __lowerCAmelCase : List[str] = 3 __lowerCAmelCase : int = 8 __lowerCAmelCase : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) __lowerCAmelCase : int = model(lowercase__ , lowercase__ ) __lowerCAmelCase : Union[str, Any] = bark_model(lowercase__ , lowercase__ ) __lowerCAmelCase : List[Any] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError('''initial and new outputs don\'t have the same shape''' ) if (output_new_model - output_old_model).abs().max().item() > 1E-3: raise ValueError('''initial and new outputs are not equal''' ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) model.save_pretrained(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = os.path.join(lowercase__ , lowercase__ ) __lowerCAmelCase : Union[str, Any] = BarkSemanticConfig.from_pretrained(os.path.join(lowercase__ , '''config.json''' ) ) __lowerCAmelCase : List[str] = BarkCoarseConfig.from_pretrained(os.path.join(lowercase__ , '''config.json''' ) ) __lowerCAmelCase : List[str] = BarkFineConfig.from_pretrained(os.path.join(lowercase__ , '''config.json''' ) ) __lowerCAmelCase : Any = EncodecConfig.from_pretrained('''facebook/encodec_24khz''' ) __lowerCAmelCase : Any = BarkSemanticModel.from_pretrained(lowercase__ ) __lowerCAmelCase : List[Any] = BarkCoarseModel.from_pretrained(lowercase__ ) __lowerCAmelCase : Optional[int] = BarkFineModel.from_pretrained(lowercase__ ) __lowerCAmelCase : Union[str, Any] = EncodecModel.from_pretrained('''facebook/encodec_24khz''' ) __lowerCAmelCase : Any = BarkConfig.from_sub_model_configs( lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) __lowerCAmelCase : Optional[Any] = BarkModel(lowercase__ ) __lowerCAmelCase : str = semantic __lowerCAmelCase : Dict = coarseAcoustic __lowerCAmelCase : int = fineAcoustic __lowerCAmelCase : Tuple = codec __lowerCAmelCase : Optional[Any] = bark_generation_config Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) bark.save_pretrained(lowercase__ , repo_id=lowercase__ , push_to_hub=lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("model_type", type=str, help="text, coarse or fine.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--is_small", action="store_true", help="convert the small version instead of the large.") _UpperCamelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
275
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
1
def _lowercase ( lowercase__ , lowercase__ , lowercase__=False ): if isinstance(lowercase__ , lowercase__ ) and isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : List[Any] = len(set_a.intersection(lowercase__ ) ) if alternative_union: __lowerCAmelCase : List[Any] = len(lowercase__ ) + len(lowercase__ ) else: __lowerCAmelCase : List[Any] = len(set_a.union(lowercase__ ) ) return intersection / union if isinstance(lowercase__ , (list, tuple) ) and isinstance(lowercase__ , (list, tuple) ): __lowerCAmelCase : str = [element for element in set_a if element in set_b] if alternative_union: __lowerCAmelCase : Any = len(lowercase__ ) + len(lowercase__ ) return len(lowercase__ ) / union else: __lowerCAmelCase : Optional[Any] = set_a + [element for element in set_b if element not in set_a] return len(lowercase__ ) / len(lowercase__ ) return len(lowercase__ ) / len(lowercase__ ) return None if __name__ == "__main__": _UpperCamelCase = {"a", "b", "c", "d", "e"} _UpperCamelCase = {"c", "d", "e", "f", "h", "i"} print(jaccard_similarity(set_a, set_b))
275
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=False , A_=True , A_="None" , A_=3 , A_=4 , A_=None , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = parent __lowerCAmelCase : List[str] = batch_size __lowerCAmelCase : Dict = seq_length __lowerCAmelCase : List[Any] = is_training __lowerCAmelCase : List[Any] = use_input_mask __lowerCAmelCase : Optional[int] = use_token_type_ids __lowerCAmelCase : Tuple = use_labels __lowerCAmelCase : str = vocab_size __lowerCAmelCase : int = hidden_size __lowerCAmelCase : Any = num_hidden_layers __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : int = hidden_act __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Union[str, Any] = type_sequence_label_size __lowerCAmelCase : Optional[int] = initializer_range __lowerCAmelCase : int = num_labels __lowerCAmelCase : int = num_choices __lowerCAmelCase : List[str] = relative_attention __lowerCAmelCase : Union[str, Any] = position_biased_input __lowerCAmelCase : int = pos_att_type __lowerCAmelCase : List[Any] = scope def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : int = None if self.use_input_mask: __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __lowerCAmelCase : List[str] = None if self.use_token_type_ids: __lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : int = None __lowerCAmelCase : List[str] = None if self.use_labels: __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_config() __lowerCAmelCase : Dict = 300 return config def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = DebertaModel(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ )[0] __lowerCAmelCase : Any = model(A_ , token_type_ids=A_ )[0] __lowerCAmelCase : List[str] = model(A_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = DebertaForMaskedLM(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = self.num_labels __lowerCAmelCase : Tuple = DebertaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Optional[int] = DebertaForTokenClassification(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Tuple = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = DebertaForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : int = model( A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ) : Tuple = config_and_inputs __lowerCAmelCase : Tuple = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowercase (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : int = DebertaModelTester(self ) __lowerCAmelCase : List[Any] = ConfigTester(self , config_class=A_ , hidden_size=37 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*A_ ) @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[int] = DebertaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowercase (unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' pass @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = DebertaModel.from_pretrained('''microsoft/deberta-base''' ) __lowerCAmelCase : Tuple = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase : Optional[int] = model(A_ , attention_mask=A_ )[0] # compare the actual values for a slice. __lowerCAmelCase : Optional[Any] = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , A_ , atol=1e-4 ) , f"""{output[:, 1:4, 1:4]}""" )
275
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def _lowercase ( lowercase__ ): return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def _lowercase ( lowercase__ ): __lowerCAmelCase : Optional[int] = create_tensor(lowercase__ ) __lowerCAmelCase : Tuple = gather(lowercase__ ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def _lowercase ( lowercase__ ): __lowerCAmelCase : List[Any] = [state.process_index] __lowerCAmelCase : str = gather_object(lowercase__ ) assert len(lowercase__ ) == state.num_processes, f"""{gathered_obj}, {len(lowercase__ )} != {state.num_processes}""" assert gathered_obj == list(range(state.num_processes ) ), f"""{gathered_obj} != {list(range(state.num_processes ) )}""" def _lowercase ( lowercase__ ): __lowerCAmelCase : Tuple = create_tensor(lowercase__ ) __lowerCAmelCase : List[Any] = broadcast(lowercase__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def _lowercase ( lowercase__ ): # We need to pad the tensor with one more element if we are the main process # to ensure that we can pad if state.is_main_process: __lowerCAmelCase : int = torch.arange(state.num_processes + 1 ).to(state.device ) else: __lowerCAmelCase : str = torch.arange(state.num_processes ).to(state.device ) __lowerCAmelCase : List[str] = pad_across_processes(lowercase__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def _lowercase ( lowercase__ ): # For now runs on only two processes if state.num_processes != 2: return __lowerCAmelCase : Optional[int] = create_tensor(lowercase__ ) __lowerCAmelCase : str = reduce(lowercase__ , '''sum''' ) __lowerCAmelCase : int = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(lowercase__ , lowercase__ ), f"""{reduced_tensor} != {truth_tensor}""" def _lowercase ( lowercase__ ): # For now runs on only two processes if state.num_processes != 2: return __lowerCAmelCase : str = create_tensor(lowercase__ ) __lowerCAmelCase : Dict = reduce(lowercase__ , '''mean''' ) __lowerCAmelCase : List[str] = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(lowercase__ , lowercase__ ), f"""{reduced_tensor} != {truth_tensor}""" def _lowercase ( lowercase__ ): # For xla_spawn (TPUs) main() def _lowercase ( ): __lowerCAmelCase : Union[str, Any] = PartialState() state.print(f"""State: {state}""" ) state.print('''testing gather''' ) test_gather(lowercase__ ) state.print('''testing gather_object''' ) test_gather_object(lowercase__ ) state.print('''testing broadcast''' ) test_broadcast(lowercase__ ) state.print('''testing pad_across_processes''' ) test_pad_across_processes(lowercase__ ) state.print('''testing reduce_sum''' ) test_reduce_sum(lowercase__ ) state.print('''testing reduce_mean''' ) test_reduce_mean(lowercase__ ) if __name__ == "__main__": main()
275
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : str = [] for rt in rc.restypes: __lowerCAmelCase : List[Any] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase : List[str] = {name: i for i, name in enumerate(lowercase__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) __lowerCAmelCase : List[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Optional[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Tuple = torch.tensor( lowercase__ , dtype=torch.floataa , device=protein['''aatype'''].device , ) __lowerCAmelCase : List[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : int = residx_atomaa_mask __lowerCAmelCase : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase : int = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase : str = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase : Optional[int] = rc.restype_atoa[restype_letter] __lowerCAmelCase : Optional[Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase : str = rc.atom_order[atom_name] __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : Any = residx_atomaa_mask return protein def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = tree_map(lambda lowercase__ : torch.tensor(lowercase__ , device=batch['''aatype'''].device ) , lowercase__ , np.ndarray ) __lowerCAmelCase : Tuple = tensor_tree_map(lambda lowercase__ : np.array(lowercase__ ) , make_atomaa_masks(lowercase__ ) ) return out
275
1
import os import sys import unittest _UpperCamelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path _UpperCamelCase = os.path.join(git_repo_path, "src", "transformers") _UpperCamelCase = "\n{0} = None\n" _UpperCamelCase = "\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n" _UpperCamelCase = "\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n" class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[int] = find_backend(''' _import_structure["models.albert"].append("AlbertTokenizerFast")''' ) self.assertIsNone(A_ ) __lowerCAmelCase : List[str] = find_backend(''' if not is_tokenizers_available():''' ) self.assertEqual(A_ , '''tokenizers''' ) __lowerCAmelCase : int = find_backend(''' if not is_tensorflow_text_available():''' ) self.assertEqual(A_ , '''tensorflow_text''' ) __lowerCAmelCase : List[Any] = find_backend(''' if not (is_sentencepiece_available() and is_tokenizers_available()):''' ) self.assertEqual(A_ , '''sentencepiece_and_tokenizers''' ) __lowerCAmelCase : List[str] = find_backend( ''' if not (is_sentencepiece_available() and is_tensorflow_text_available()):''' ) self.assertEqual(A_ , '''sentencepiece_and_tensorflow_text''' ) __lowerCAmelCase : List[str] = find_backend( ''' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):''' ) self.assertEqual(A_ , '''sentencepiece_and_tokenizers_and_vision''' ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('''torch''' , A_ ) self.assertIn('''tensorflow_text''' , A_ ) self.assertIn('''sentencepiece_and_tokenizers''' , A_ ) # Likewise, we can't assert on the exact content of a key self.assertIn('''BertModel''' , objects['''torch'''] ) self.assertIn('''TFBertModel''' , objects['''tf'''] ) self.assertIn('''FlaxBertModel''' , objects['''flax'''] ) self.assertIn('''BertModel''' , objects['''torch'''] ) self.assertIn('''TFBertTokenizer''' , objects['''tensorflow_text'''] ) self.assertIn('''convert_slow_tokenizer''' , objects['''sentencepiece_and_tokenizers'''] ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = create_dummy_object('''CONSTANT''' , '''\'torch\'''' ) self.assertEqual(A_ , '''\nCONSTANT = None\n''' ) __lowerCAmelCase : int = create_dummy_object('''function''' , '''\'torch\'''' ) self.assertEqual( A_ , '''\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n''' ) __lowerCAmelCase : List[Any] = ''' class FakeClass(metaclass=DummyObject): _backends = \'torch\' def __init__(self, *args, **kwargs): requires_backends(self, \'torch\') ''' __lowerCAmelCase : int = create_dummy_object('''FakeClass''' , '''\'torch\'''' ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Tuple = '''# This file is autogenerated by the command `make fix-copies`, do not edit. from ..utils import DummyObject, requires_backends CONSTANT = None def function(*args, **kwargs): requires_backends(function, ["torch"]) class FakeClass(metaclass=DummyObject): _backends = ["torch"] def __init__(self, *args, **kwargs): requires_backends(self, ["torch"]) ''' __lowerCAmelCase : Tuple = create_dummy_files({'''torch''': ['''CONSTANT''', '''function''', '''FakeClass''']} ) self.assertEqual(dummy_files['''torch'''] , A_ )
275
def _lowercase ( lowercase__ ): if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) __lowerCAmelCase : int = sorted(string.lower() ) return len(lowercase__ ) == len(set(lowercase__ ) ) if __name__ == "__main__": _UpperCamelCase = input("Enter a string ").strip() _UpperCamelCase = is_isogram(input_str) print(F"{input_str} is {'an' if isogram else 'not an'} isogram.")
275
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _UpperCamelCase = {"configuration_yolos": ["YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP", "YolosConfig", "YolosOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["YolosFeatureExtractor"] _UpperCamelCase = ["YolosImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ "YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST", "YolosForObjectDetection", "YolosModel", "YolosPreTrainedModel", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
1
import pytest import datasets # Import fixture modules as plugins _UpperCamelCase = ["tests.fixtures.files", "tests.fixtures.hub", "tests.fixtures.fsspec"] def _lowercase ( lowercase__ , lowercase__ ): # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def _lowercase ( lowercase__ ): config.addinivalue_line('''markers''' , '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? __lowerCAmelCase : int = tmp_path_factory.getbasetemp() / '''cache''' __lowerCAmelCase : Optional[Any] = test_hf_cache_home / '''datasets''' __lowerCAmelCase : Tuple = test_hf_cache_home / '''metrics''' __lowerCAmelCase : Optional[int] = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''' , str(lowercase__ ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''' , str(lowercase__ ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''' , str(lowercase__ ) ) __lowerCAmelCase : List[str] = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''' , str(lowercase__ ) ) __lowerCAmelCase : List[str] = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(lowercase__ ) ) @pytest.fixture(autouse=lowercase__ , scope='''session''' ) def _lowercase ( ): datasets.disable_progress_bar() @pytest.fixture(autouse=lowercase__ ) def _lowercase ( lowercase__ ): # don't take tests into account when counting downloads monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''' , lowercase__ ) @pytest.fixture def _lowercase ( lowercase__ ): # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''' , lowercase__ )
275
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = np.full((len(lowercase__ ), sequence_length, 2) , lowercase__ ) else: __lowerCAmelCase : Optional[int] = np.full((len(lowercase__ ), sequence_length) , lowercase__ ) for i, tensor in enumerate(lowercase__ ): if padding_side == "right": if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : int = tensor[:sequence_length] else: if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = ord(lowercase__ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True __lowerCAmelCase : int = unicodedata.category(lowercase__ ) if cat.startswith('''P''' ): return True return False @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = True _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = -100 _UpperCamelCase = "pt" def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCAmelCase : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCAmelCase : List[Any] = self.tokenizer.pad( A_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCAmelCase : Dict = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCAmelCase : Optional[int] = self.tokenizer.padding_side if padding_side == "right": __lowerCAmelCase : Any = [ list(A_ ) + [self.label_pad_token_id] * (sequence_length - len(A_ )) for label in labels ] else: __lowerCAmelCase : Optional[int] = [ [self.label_pad_token_id] * (sequence_length - len(A_ )) + list(A_ ) for label in labels ] __lowerCAmelCase : Tuple = [feature['''ner_tags'''] for feature in features] __lowerCAmelCase : List[Any] = padding_tensor(A_ , -1 , A_ , A_ ) __lowerCAmelCase : Optional[int] = [feature['''original_entity_spans'''] for feature in features] __lowerCAmelCase : Any = padding_tensor(A_ , (-1, -1) , A_ , A_ ) __lowerCAmelCase : Optional[Any] = {k: torch.tensor(A_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
275
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "uw-madison/mra-base-512-4": "https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json", } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """mra""" def __init__( self , A_=5_0265 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=1 , A_=0.02 , A_=1e-5 , A_="absolute" , A_=4 , A_="full" , A_=0 , A_=0 , A_=1 , A_=0 , A_=2 , **A_ , ) ->Optional[int]: '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) __lowerCAmelCase : Union[str, Any] = vocab_size __lowerCAmelCase : Union[str, Any] = max_position_embeddings __lowerCAmelCase : List[Any] = hidden_size __lowerCAmelCase : Any = num_hidden_layers __lowerCAmelCase : Optional[int] = num_attention_heads __lowerCAmelCase : Optional[int] = intermediate_size __lowerCAmelCase : Optional[Any] = hidden_act __lowerCAmelCase : Tuple = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : List[Any] = type_vocab_size __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : List[str] = position_embedding_type __lowerCAmelCase : Tuple = block_per_row __lowerCAmelCase : List[Any] = approx_mode __lowerCAmelCase : str = initial_prior_first_n_blocks __lowerCAmelCase : Optional[Any] = initial_prior_diagonal_n_blocks
275
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase (unittest.TestCase ): _UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Tuple = (3, 32, 128) __lowerCAmelCase : List[str] = tempfile.mkdtemp() # fmt: off __lowerCAmelCase : List[str] = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on __lowerCAmelCase : Optional[int] = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) __lowerCAmelCase : Union[str, Any] = { '''do_normalize''': False, '''do_resize''': True, '''image_processor_type''': '''ViTImageProcessor''', '''resample''': 3, '''size''': {'''height''': 32, '''width''': 128}, } __lowerCAmelCase : Optional[Any] = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(A_ , A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) __lowerCAmelCase : str = Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) return image_input def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Union[str, Any] = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : int = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) __lowerCAmelCase : int = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.get_image_processor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Optional[int] = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = image_processor(A_ , return_tensors='''np''' ) __lowerCAmelCase : Tuple = processor(images=A_ , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Union[str, Any] = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Any = '''test''' __lowerCAmelCase : Dict = processor(text=A_ ) __lowerCAmelCase : str = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : str = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = '''test''' __lowerCAmelCase : int = self.prepare_image_inputs() __lowerCAmelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''pixel_values''', '''labels'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Optional[int] = processor.char_decode(A_ ) __lowerCAmelCase : Tuple = tokenizer.batch_decode(A_ ) __lowerCAmelCase : Any = [seq.replace(''' ''' , '''''' ) for seq in decoded_tok] self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = self.prepare_image_inputs() __lowerCAmelCase : List[Any] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_image_processor() __lowerCAmelCase : List[str] = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = torch.randn(1 , 27 , 38 ) __lowerCAmelCase : Optional[int] = torch.randn(1 , 27 , 5_0257 ) __lowerCAmelCase : Optional[Any] = torch.randn(1 , 27 , 3_0522 ) __lowerCAmelCase : List[str] = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['''generated_text''', '''scores''', '''char_preds''', '''bpe_preds''', '''wp_preds'''] )
275
1
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() __lowerCAmelCase : Tuple = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : str = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } __lowerCAmelCase : Dict = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_6000, '''return_attention_mask''': False, '''do_normalize''': True, } __lowerCAmelCase : int = tempfile.mkdtemp() __lowerCAmelCase : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCAmelCase : List[str] = os.path.join(self.tmpdirname , A_ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) # load decoder from hub __lowerCAmelCase : Dict = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase__ ( self , **A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = self.add_kwargs_tokens_map.copy() kwargs.update(A_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->str: '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_feature_extractor() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : List[str] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Dict = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match __lowerCAmelCase : List[str] = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(A_ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=A_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.get_feature_extractor() __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : Optional[int] = self.get_decoder() __lowerCAmelCase : Any = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : int = floats_list((3, 1000) ) __lowerCAmelCase : Tuple = feature_extractor(A_ , return_tensors='''np''' ) __lowerCAmelCase : str = processor(A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = self.get_feature_extractor() __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : Dict = '''This is a test string''' __lowerCAmelCase : Dict = processor(text=A_ ) __lowerCAmelCase : List[Any] = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self , A_=(2, 10, 16) , A_=77 ) ->Tuple: '''simple docstring''' np.random.seed(A_ ) return np.random.rand(*A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.get_feature_extractor() __lowerCAmelCase : Optional[int] = self.get_tokenizer() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : int = self._get_dummy_logits(shape=(10, 16) , seed=13 ) __lowerCAmelCase : Any = processor.decode(A_ ) __lowerCAmelCase : str = decoder.decode_beams(A_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase__ ( self , A_ ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : int = self.get_feature_extractor() __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : Any = self.get_decoder() __lowerCAmelCase : Any = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: __lowerCAmelCase : Union[str, Any] = processor.batch_decode(A_ ) else: with get_context(A_ ).Pool() as pool: __lowerCAmelCase : Optional[int] = processor.batch_decode(A_ , A_ ) __lowerCAmelCase : List[str] = list(A_ ) with get_context('''fork''' ).Pool() as p: __lowerCAmelCase : str = decoder.decode_beams_batch(A_ , A_ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A_ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(A_ , decoded_processor.logit_score ) self.assertListEqual(A_ , decoded_processor.lm_score ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = self.get_feature_extractor() __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : List[Any] = self._get_dummy_logits() __lowerCAmelCase : Tuple = 15 __lowerCAmelCase : Any = -20.0 __lowerCAmelCase : List[str] = -4.0 __lowerCAmelCase : Optional[int] = processor.batch_decode( A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) __lowerCAmelCase : Union[str, Any] = decoded_processor_out.text __lowerCAmelCase : Union[str, Any] = list(A_ ) with get_context('''fork''' ).Pool() as pool: __lowerCAmelCase : List[str] = decoder.decode_beams_batch( A_ , A_ , beam_width=A_ , beam_prune_logp=A_ , token_min_logp=A_ , ) __lowerCAmelCase : Any = [d[0][0] for d in decoded_decoder_out] __lowerCAmelCase : Optional[Any] = [d[0][2] for d in decoded_decoder_out] __lowerCAmelCase : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , A_ ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , A_ , atol=1e-3 ) ) self.assertTrue(np.array_equal(A_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = self.get_feature_extractor() __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : List[Any] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) __lowerCAmelCase : List[str] = self._get_dummy_logits() __lowerCAmelCase : str = 2.0 __lowerCAmelCase : int = 5.0 __lowerCAmelCase : Any = -20.0 __lowerCAmelCase : Union[str, Any] = True __lowerCAmelCase : Optional[int] = processor.batch_decode( A_ , alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) __lowerCAmelCase : Dict = decoded_processor_out.text __lowerCAmelCase : Optional[int] = list(A_ ) decoder.reset_params( alpha=A_ , beta=A_ , unk_score_offset=A_ , lm_score_boundary=A_ , ) with get_context('''fork''' ).Pool() as pool: __lowerCAmelCase : int = decoder.decode_beams_batch( A_ , A_ , ) __lowerCAmelCase : int = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A_ , A_ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , A_ ) __lowerCAmelCase : List[Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : Optional[int] = processor.decoder.model_container[processor.decoder._model_key] __lowerCAmelCase : int = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() __lowerCAmelCase : Union[str, Any] = os.listdir(A_ ) __lowerCAmelCase : Dict = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = snapshot_download('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained(A_ ) __lowerCAmelCase : List[str] = processor.decoder.model_container[processor.decoder._model_key] __lowerCAmelCase : List[Any] = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() __lowerCAmelCase : List[str] = os.listdir(A_ ) __lowerCAmelCase : Union[str, Any] = os.listdir(A_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : Optional[int] = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : str = floats_list((3, 1000) ) __lowerCAmelCase : int = processor_wavaveca(A_ , return_tensors='''np''' ) __lowerCAmelCase : int = processor_auto(A_ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) __lowerCAmelCase : Optional[Any] = self._get_dummy_logits() __lowerCAmelCase : int = processor_wavaveca.batch_decode(A_ ) __lowerCAmelCase : Tuple = processor_auto.batch_decode(A_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : int = self.get_feature_extractor() __lowerCAmelCase : Optional[int] = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_decoder() __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A_ , feature_extractor=A_ , decoder=A_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase__ ( A_ , A_ ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[str] = [d[key] for d in offsets] return retrieved_list def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : Optional[Any] = self._get_dummy_logits()[0] __lowerCAmelCase : Any = processor.decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __lowerCAmelCase : Any = self._get_dummy_logits() __lowerCAmelCase : str = processor.batch_decode(A_ , output_word_offsets=A_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(A_ , A_ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : int = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=A_ ) __lowerCAmelCase : Union[str, Any] = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_6000 ) ) __lowerCAmelCase : Optional[int] = iter(A_ ) __lowerCAmelCase : int = next(A_ ) __lowerCAmelCase : str = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) __lowerCAmelCase : str = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train __lowerCAmelCase : Tuple = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): __lowerCAmelCase : Any = model(A_ ).logits.cpu().numpy() __lowerCAmelCase : Dict = processor.decode(logits[0] , output_word_offsets=A_ ) __lowerCAmelCase : Any = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate __lowerCAmelCase : Union[str, Any] = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] __lowerCAmelCase : Union[str, Any] = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , A_ ) self.assertEqual(''' '''.join(self.get_from_offsets(A_ , '''word''' ) ) , output.text ) # output times __lowerCAmelCase : Tuple = torch.tensor(self.get_from_offsets(A_ , '''start_time''' ) ) __lowerCAmelCase : Tuple = torch.tensor(self.get_from_offsets(A_ , '''end_time''' ) ) # fmt: off __lowerCAmelCase : Tuple = torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) __lowerCAmelCase : Dict = torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=0.01 ) )
275
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = self.dummy_uncond_unet __lowerCAmelCase : Any = PNDMScheduler() __lowerCAmelCase : Dict = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' ).images __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''google/ddpm-cifar10-32''' __lowerCAmelCase : Union[str, Any] = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : int = PNDMScheduler() __lowerCAmelCase : Any = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
1
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand _UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def _lowercase ( lowercase__ ): if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(lowercase__ ): return ext raise Exception( f"""Unable to determine file format from file extension {path}. """ f"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def _lowercase ( lowercase__ ): __lowerCAmelCase : Any = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) __lowerCAmelCase : Dict = try_infer_format_from_ext(args.input ) if args.format == '''infer''' else args.format __lowerCAmelCase : Dict = PipelineDataFormat.from_str( format=lowercase__ , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(lowercase__ , lowercase__ ) class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = nlp __lowerCAmelCase : Tuple = reader @staticmethod def UpperCamelCase__ ( A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : str = parser.add_parser('''run''' , help='''Run a pipeline through the CLI''' ) run_parser.add_argument('''--task''' , choices=get_supported_tasks() , help='''Task to run''' ) run_parser.add_argument('''--input''' , type=A_ , help='''Path to the file to use for inference''' ) run_parser.add_argument('''--output''' , type=A_ , help='''Path to the file that will be used post to write results.''' ) run_parser.add_argument('''--model''' , type=A_ , help='''Name or path to the model to instantiate.''' ) run_parser.add_argument('''--config''' , type=A_ , help='''Name or path to the model\'s config to instantiate.''' ) run_parser.add_argument( '''--tokenizer''' , type=A_ , help='''Name of the tokenizer to use. (default: same as the model name)''' ) run_parser.add_argument( '''--column''' , type=A_ , help='''Name of the column to use as input. (For multi columns input as QA use column1,columns2)''' , ) run_parser.add_argument( '''--format''' , type=A_ , default='''infer''' , choices=PipelineDataFormat.SUPPORTED_FORMATS , help='''Input format to read from''' , ) run_parser.add_argument( '''--device''' , type=A_ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) run_parser.add_argument('''--overwrite''' , action='''store_true''' , help='''Allow overwriting the output file.''' ) run_parser.set_defaults(func=A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Any = self._nlp, [] for entry in self._reader: __lowerCAmelCase : Optional[int] = nlp(**A_ ) if self._reader.is_multi_columns else nlp(A_ ) if isinstance(A_ , A_ ): outputs.append(A_ ) else: outputs += output # Saving data if self._nlp.binary_output: __lowerCAmelCase : Union[str, Any] = self._reader.save_binary(A_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(A_ )
275
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
1
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def _lowercase ( lowercase__ ): __lowerCAmelCase : List[Any] = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ ): __lowerCAmelCase : int = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: __lowerCAmelCase : List[str] = s_dict.pop(lowercase__ ) elif "subsample" in key: __lowerCAmelCase : Any = s_dict.pop(lowercase__ ) def _lowercase ( lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : List[Any] = emb.weight.shape __lowerCAmelCase : int = nn.Linear(lowercase__ , lowercase__ , bias=lowercase__ ) __lowerCAmelCase : str = emb.weight.data return lin_layer def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[Any] = torch.load(lowercase__ , map_location='''cpu''' ) __lowerCAmelCase : int = mam_aaa['''args'''] __lowerCAmelCase : Optional[Any] = mam_aaa['''model'''] __lowerCAmelCase : List[str] = state_dict['''decoder.output_projection.weight'''] remove_ignore_keys_(lowercase__ ) rename_keys(lowercase__ ) __lowerCAmelCase : Any = state_dict['''decoder.embed_tokens.weight'''].shape[0] __lowerCAmelCase : List[str] = args.share_decoder_input_output_embed __lowerCAmelCase : Any = [int(lowercase__ ) for i in args.conv_kernel_sizes.split(''',''' )] __lowerCAmelCase : Union[str, Any] = SpeechaTextConfig( vocab_size=lowercase__ , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , num_conv_layers=len(lowercase__ ) , conv_channels=args.conv_channels , conv_kernel_sizes=lowercase__ , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=lowercase__ , num_beams=5 , max_length=2_0_0 , use_cache=lowercase__ , decoder_start_token_id=2 , early_stopping=lowercase__ , ) __lowerCAmelCase : Tuple = SpeechaTextForConditionalGeneration(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : Tuple = model.model.load_state_dict(lowercase__ , strict=lowercase__ ) if len(lowercase__ ) > 0 and not set(lowercase__ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: __lowerCAmelCase : int = make_linear_from_emb(model.model.decoder.embed_tokens ) else: __lowerCAmelCase : str = lm_head_weights model.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("--fairseq_path", type=str, help="Path to the fairseq model (.pt) file.") parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") _UpperCamelCase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
275
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _UpperCamelCase = { "AI-Sweden/gpt-sw3-126m": 2048, "AI-Sweden/gpt-sw3-350m": 2048, "AI-Sweden/gpt-sw3-1.6b": 2048, "AI-Sweden/gpt-sw3-6.7b": 2048, "AI-Sweden/gpt-sw3-20b": 2048, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_ , A_=False , A_=False , A_=False , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCAmelCase : int = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCAmelCase : Union[str, Any] = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCAmelCase : str = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCAmelCase : Any = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token __lowerCAmelCase : int = eos_token if bos_token is None else bos_token else: __lowerCAmelCase : Optional[int] = '''<pad>''' if pad_token is None else pad_token __lowerCAmelCase : List[str] = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=A_ , remove_space=A_ , keep_accents=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Union[str, Any] = do_lower_case __lowerCAmelCase : Union[str, Any] = remove_space __lowerCAmelCase : int = keep_accents __lowerCAmelCase : Union[str, Any] = vocab_file __lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # Used for whitespace normalization in input texts # fmt : off __lowerCAmelCase : List[Any] = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCAmelCase : int = re.compile( f"""[{"".join(map(A_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.__dict__.copy() __lowerCAmelCase : List[Any] = None return state def __setstate__( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.non_printing_characters_re.sub('''''' , A_ ) # Normalize whitespaces __lowerCAmelCase : List[str] = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCAmelCase : Tuple = unicodedata.normalize('''NFC''' , A_ ) return text def UpperCamelCase__ ( self , A_ , **A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = self.preprocess_text(A_ ) return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' return self.sp_model.PieceToId(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.IdToPiece(A_ ) @staticmethod def UpperCamelCase__ ( A_ ) ->str: '''simple docstring''' return out_string def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = [] __lowerCAmelCase : Tuple = '''''' __lowerCAmelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string def UpperCamelCase__ ( self ) ->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : str = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Any = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = False ) ->Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[Any] = self.preprocess_text(A_ ) __lowerCAmelCase : Dict = self.sp_model.encode(A_ ) else: __lowerCAmelCase : Dict = [self.preprocess_text(A_ ) for t in text] __lowerCAmelCase : Optional[int] = self.sp_model.encode(A_ ) if return_tensors is True or return_tensors == "pt": __lowerCAmelCase : Tuple = torch.tensor(A_ ) return token_ids def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.decode(A_ ) def UpperCamelCase__ ( self , A_ ) ->List[int]: '''simple docstring''' __lowerCAmelCase : int = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCAmelCase : Any = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(A_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=A_ )
275
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "unc-nlp/lxmert-base-uncased": "https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json", } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """lxmert""" _UpperCamelCase = {} def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=9500 , A_=1600 , A_=400 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=9 , A_=5 , A_=5 , A_=2048 , A_=4 , A_=6.67 , A_=True , A_=True , A_=True , A_=True , A_=True , A_=True , A_=True , **A_ , ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = vocab_size __lowerCAmelCase : List[str] = hidden_size __lowerCAmelCase : str = num_attention_heads __lowerCAmelCase : Any = hidden_act __lowerCAmelCase : Optional[Any] = intermediate_size __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Tuple = max_position_embeddings __lowerCAmelCase : List[Any] = type_vocab_size __lowerCAmelCase : Any = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = num_qa_labels __lowerCAmelCase : str = num_object_labels __lowerCAmelCase : Optional[int] = num_attr_labels __lowerCAmelCase : Optional[Any] = l_layers __lowerCAmelCase : int = x_layers __lowerCAmelCase : Dict = r_layers __lowerCAmelCase : Union[str, Any] = visual_feat_dim __lowerCAmelCase : Optional[int] = visual_pos_dim __lowerCAmelCase : Dict = visual_loss_normalizer __lowerCAmelCase : str = task_matched __lowerCAmelCase : int = task_mask_lm __lowerCAmelCase : int = task_obj_predict __lowerCAmelCase : int = task_qa __lowerCAmelCase : Tuple = visual_obj_loss __lowerCAmelCase : Union[str, Any] = visual_attr_loss __lowerCAmelCase : str = visual_feat_loss __lowerCAmelCase : Union[str, Any] = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**A_ )
275
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """table-transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) ->Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowerCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(A_ , A_ ): __lowerCAmelCase : int = backbone_config.get('''model_type''' ) __lowerCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __lowerCAmelCase : Any = config_class.from_dict(A_ ) # set timm attributes to None __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = None, None, None __lowerCAmelCase : Tuple = use_timm_backbone __lowerCAmelCase : Optional[Any] = backbone_config __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = num_queries __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_ffn_dim __lowerCAmelCase : Optional[int] = encoder_layers __lowerCAmelCase : List[str] = encoder_attention_heads __lowerCAmelCase : str = decoder_ffn_dim __lowerCAmelCase : Union[str, Any] = decoder_layers __lowerCAmelCase : Any = decoder_attention_heads __lowerCAmelCase : Optional[int] = dropout __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : Tuple = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : List[str] = init_std __lowerCAmelCase : Tuple = init_xavier_std __lowerCAmelCase : Any = encoder_layerdrop __lowerCAmelCase : List[Any] = decoder_layerdrop __lowerCAmelCase : Optional[Any] = encoder_layers __lowerCAmelCase : Optional[Any] = auxiliary_loss __lowerCAmelCase : Optional[Any] = position_embedding_type __lowerCAmelCase : Tuple = backbone __lowerCAmelCase : Any = use_pretrained_backbone __lowerCAmelCase : int = dilation # Hungarian matcher __lowerCAmelCase : Dict = class_cost __lowerCAmelCase : List[str] = bbox_cost __lowerCAmelCase : int = giou_cost # Loss coefficients __lowerCAmelCase : Optional[Any] = mask_loss_coefficient __lowerCAmelCase : Tuple = dice_loss_coefficient __lowerCAmelCase : int = bbox_loss_coefficient __lowerCAmelCase : List[Any] = giou_loss_coefficient __lowerCAmelCase : int = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.d_model class __lowercase (_UpperCAmelCase ): _UpperCamelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def UpperCamelCase__ ( self ) ->float: '''simple docstring''' return 1e-5 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 12
275
1
from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image _UpperCamelCase = ["text", "image", "audio"] def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((5_1_2, 5_1_2) ) ) elif input_type == "audio": inputs.append(torch.ones(3_0_0_0 ) ) elif isinstance(lowercase__ , lowercase__ ): inputs.append(create_inputs(lowercase__ ) ) else: raise ValueError(f"""Invalid type requested: {input_type}""" ) return inputs def _lowercase ( lowercase__ ): __lowerCAmelCase : Any = [] for output in outputs: if isinstance(lowercase__ , (str, AgentText) ): output_types.append('''text''' ) elif isinstance(lowercase__ , (Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(lowercase__ , (torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(f"""Invalid output: {output}""" ) return output_types @is_tool_test class __lowercase : def UpperCamelCase__ ( self ) ->str: '''simple docstring''' self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) __lowerCAmelCase : str = self.tool.inputs for _input in inputs: if isinstance(_input , A_ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) __lowerCAmelCase : Optional[Any] = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = create_inputs(self.tool.inputs ) __lowerCAmelCase : Dict = self.tool(*A_ ) # There is a single output if len(self.tool.outputs ) == 1: __lowerCAmelCase : str = [outputs] self.assertListEqual(output_types(A_ ) , self.tool.outputs ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : List[Any] = create_inputs(self.tool.inputs ) __lowerCAmelCase : Dict = self.tool(*A_ ) if not isinstance(A_ , A_ ): __lowerCAmelCase : List[Any] = [outputs] self.assertEqual(len(A_ ) , len(self.tool.outputs ) ) for output, output_type in zip(A_ , self.tool.outputs ): __lowerCAmelCase : Union[str, Any] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = create_inputs(self.tool.inputs ) __lowerCAmelCase : List[str] = [] for _input, input_type in zip(A_ , self.tool.inputs ): if isinstance(A_ , A_ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error __lowerCAmelCase : Dict = self.tool(*A_ ) if not isinstance(A_ , A_ ): __lowerCAmelCase : Any = [outputs] self.assertEqual(len(A_ ) , len(self.tool.outputs ) )
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """git_vision_model""" def __init__( self , A_=768 , A_=3072 , A_=12 , A_=12 , A_=3 , A_=224 , A_=16 , A_="quick_gelu" , A_=1e-5 , A_=0.0 , A_=0.02 , **A_ , ) ->Any: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Optional[Any] = hidden_size __lowerCAmelCase : Any = intermediate_size __lowerCAmelCase : Tuple = num_hidden_layers __lowerCAmelCase : Optional[int] = num_attention_heads __lowerCAmelCase : int = num_channels __lowerCAmelCase : Optional[int] = patch_size __lowerCAmelCase : Dict = image_size __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : int = layer_norm_eps __lowerCAmelCase : Optional[int] = hidden_act @classmethod def UpperCamelCase__ ( cls , A_ , **A_ ) ->"PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(A_ ) __lowerCAmelCase, __lowerCAmelCase : Any = cls.get_config_dict(A_ , **A_ ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": __lowerCAmelCase : Optional[Any] = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(A_ , **A_ ) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """git""" def __init__( self , A_=None , A_=3_0522 , A_=768 , A_=6 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=1024 , A_=0.02 , A_=1e-12 , A_=0 , A_="absolute" , A_=True , A_=False , A_=101 , A_=102 , A_=None , **A_ , ) ->Optional[int]: '''simple docstring''' super().__init__(bos_token_id=A_ , eos_token_id=A_ , pad_token_id=A_ , **A_ ) if vision_config is None: __lowerCAmelCase : Any = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) __lowerCAmelCase : Dict = GitVisionConfig(**A_ ) __lowerCAmelCase : int = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : Tuple = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Dict = hidden_act __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : Tuple = hidden_dropout_prob __lowerCAmelCase : Dict = attention_probs_dropout_prob __lowerCAmelCase : Union[str, Any] = max_position_embeddings __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Optional[int] = position_embedding_type __lowerCAmelCase : List[Any] = use_cache __lowerCAmelCase : Union[str, Any] = tie_word_embeddings __lowerCAmelCase : int = num_image_with_embedding __lowerCAmelCase : Union[str, Any] = bos_token_id __lowerCAmelCase : Optional[int] = eos_token_id def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = copy.deepcopy(self.__dict__ ) __lowerCAmelCase : Any = self.vision_config.to_dict() __lowerCAmelCase : List[Any] = self.__class__.model_type return output
275
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = {} with open(lowercase__ , '''r''' ) as file: for line_number, line in enumerate(lowercase__ ): __lowerCAmelCase : Any = line.strip() if line: __lowerCAmelCase : Dict = line.split() __lowerCAmelCase : str = line_number __lowerCAmelCase : List[str] = words[0] __lowerCAmelCase : Any = value return result def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[Any] = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : Tuple = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : List[Any] = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : str = getattr(lowercase__ , lowercase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCAmelCase : Dict = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : str = shape_pointer.shape # let's reduce dimension __lowerCAmelCase : Any = value[0] else: __lowerCAmelCase : str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __lowerCAmelCase : List[str] = value elif weight_type == "weight_v": __lowerCAmelCase : int = value elif weight_type == "bias": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : int = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : Tuple = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCAmelCase : List[str] = '''.'''.join([key, hf_param_name] ) else: __lowerCAmelCase : Optional[int] = key __lowerCAmelCase : Union[str, Any] = value if '''lm_head''' in full_key else value[0] _UpperCamelCase = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None ): __lowerCAmelCase : Any = False for key, mapped_key in MAPPING.items(): __lowerCAmelCase : Tuple = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : List[str] = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : List[Any] = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : Any = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCAmelCase : int = '''weight''' else: __lowerCAmelCase : Any = None if hf_dict is not None: rename_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return is_used return is_used def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[Any] = fairseq_model.state_dict() __lowerCAmelCase : Tuple = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : int = True else: __lowerCAmelCase : Dict = load_wavaveca_layer(lowercase__ , lowercase__ , lowercase__ ) if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : List[str] = name.split('''.''' ) __lowerCAmelCase : Any = int(items[0] ) __lowerCAmelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=True , lowercase__=False ): if config_path is not None: __lowerCAmelCase : Union[str, Any] = WavaVecaConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Optional[int] = WavaVecaConfig() if is_seq_class: __lowerCAmelCase : Optional[Any] = read_txt_into_dict(lowercase__ ) __lowerCAmelCase : int = idalabel __lowerCAmelCase : Optional[int] = WavaVecaForSequenceClassification(lowercase__ ) __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) feature_extractor.save_pretrained(lowercase__ ) elif is_finetuned: if dict_path: __lowerCAmelCase : List[str] = Dictionary.load(lowercase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCAmelCase : List[Any] = target_dict.pad_index __lowerCAmelCase : List[Any] = target_dict.bos_index __lowerCAmelCase : Optional[int] = target_dict.eos_index __lowerCAmelCase : Any = len(target_dict.symbols ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''vocab.json''' ) if not os.path.isdir(lowercase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(lowercase__ ) ) return os.makedirs(lowercase__ , exist_ok=lowercase__ ) __lowerCAmelCase : Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : int = 1 with open(lowercase__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = WavaVecaCTCTokenizer( lowercase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=lowercase__ , ) __lowerCAmelCase : List[str] = True if config.feat_extract_norm == '''layer''' else False __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=lowercase__ , tokenizer=lowercase__ ) processor.save_pretrained(lowercase__ ) __lowerCAmelCase : str = WavaVecaForCTC(lowercase__ ) else: __lowerCAmelCase : Any = WavaVecaForPreTraining(lowercase__ ) if is_finetuned or is_seq_class: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCAmelCase : Union[str, Any] = argparse.Namespace(task='''audio_pretraining''' ) __lowerCAmelCase : str = fairseq.tasks.setup_task(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowercase__ ) __lowerCAmelCase : int = model[0].eval() recursively_load_weights(lowercase__ , lowercase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
275
1
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : str = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.dummy_uncond_unet __lowerCAmelCase : int = KarrasVeScheduler() __lowerCAmelCase : Any = KarrasVePipeline(unet=A_ , scheduler=A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pipe(num_inference_steps=2 , generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : str = pipe(num_inference_steps=2 , generator=A_ , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[int] = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = '''google/ncsnpp-celebahq-256''' __lowerCAmelCase : int = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : Union[str, Any] = KarrasVeScheduler() __lowerCAmelCase : Optional[int] = KarrasVePipeline(unet=A_ , scheduler=A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Tuple = pipe(num_inference_steps=20 , generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __lowerCAmelCase : str = np.array([0.578, 0.5_811, 0.5_924, 0.5_809, 0.587, 0.5_886, 0.5_861, 0.5_802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def _lowercase ( ): __lowerCAmelCase : Any = HfArgumentParser(lowercase__ ) __lowerCAmelCase : Any = parser.parse_args_into_dataclasses()[0] __lowerCAmelCase : Union[str, Any] = TensorFlowBenchmark(args=lowercase__ ) try: __lowerCAmelCase : Tuple = parser.parse_args_into_dataclasses()[0] except ValueError as e: __lowerCAmelCase : str = '''Arg --no_{0} is no longer used, please use --no-{0} instead.''' __lowerCAmelCase : Optional[int] = ''' '''.join(str(lowercase__ ).split(''' ''' )[:-1] ) __lowerCAmelCase : List[Any] = '''''' __lowerCAmelCase : Tuple = eval(str(lowercase__ ).split(''' ''' )[-1] ) __lowerCAmelCase : Optional[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(lowercase__ ) if len(lowercase__ ) > 0: __lowerCAmelCase : Union[str, Any] = full_error_msg + begin_error_msg + str(lowercase__ ) raise ValueError(lowercase__ ) benchmark.run() if __name__ == "__main__": main()
275
def _lowercase ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __lowerCAmelCase : int = str(bin(lowercase__ ) )[2:] # remove the leading "0b" __lowerCAmelCase : Any = str(bin(lowercase__ ) )[2:] __lowerCAmelCase : List[str] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('''1''' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
275
1
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[int] = old_name if "patch_embed" in old_name: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Tuple = old_name.split('''.''' ) if layer == "0": __lowerCAmelCase : int = old_name.replace('''0''' , '''convolution1''' ) elif layer == "1": __lowerCAmelCase : Union[str, Any] = old_name.replace('''1''' , '''batchnorm_before''' ) elif layer == "3": __lowerCAmelCase : List[str] = old_name.replace('''3''' , '''convolution2''' ) else: __lowerCAmelCase : Optional[int] = old_name.replace('''4''' , '''batchnorm_after''' ) if "network" in old_name and re.search(r'''\d\.\d''' , lowercase__ ): __lowerCAmelCase : List[Any] = r'''\b\d{2}\b''' if bool(re.search(lowercase__ , lowercase__ ) ): __lowerCAmelCase : Dict = re.search(r'''\d\.\d\d.''' , lowercase__ ).group() else: __lowerCAmelCase : Optional[Any] = re.search(r'''\d\.\d.''' , lowercase__ ).group() if int(match[0] ) < 6: __lowerCAmelCase : Optional[Any] = old_name.replace(lowercase__ , '''''' ) __lowerCAmelCase : Union[str, Any] = trimmed_name.replace('''network''' , match[0] + '''.meta4D_layers.blocks.''' + match[2:-1] ) __lowerCAmelCase : int = '''intermediate_stages.''' + trimmed_name else: __lowerCAmelCase : Optional[int] = old_name.replace(lowercase__ , '''''' ) if int(match[2] ) < num_meta4D_last_stage: __lowerCAmelCase : List[str] = trimmed_name.replace('''network''' , '''meta4D_layers.blocks.''' + match[2] ) else: __lowerCAmelCase : Union[str, Any] = str(int(match[2] ) - num_meta4D_last_stage ) __lowerCAmelCase : Union[str, Any] = trimmed_name.replace('''network''' , '''meta3D_layers.blocks.''' + layer_index ) if "norm1" in old_name: __lowerCAmelCase : Optional[Any] = trimmed_name.replace('''norm1''' , '''layernorm1''' ) elif "norm2" in old_name: __lowerCAmelCase : int = trimmed_name.replace('''norm2''' , '''layernorm2''' ) elif "fc1" in old_name: __lowerCAmelCase : Tuple = trimmed_name.replace('''fc1''' , '''linear_in''' ) elif "fc2" in old_name: __lowerCAmelCase : Optional[int] = trimmed_name.replace('''fc2''' , '''linear_out''' ) __lowerCAmelCase : str = '''last_stage.''' + trimmed_name elif "network" in old_name and re.search(r'''.\d.''' , lowercase__ ): __lowerCAmelCase : Any = old_name.replace('''network''' , '''intermediate_stages''' ) if "fc" in new_name: __lowerCAmelCase : Union[str, Any] = new_name.replace('''fc''' , '''convolution''' ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): __lowerCAmelCase : Any = new_name.replace('''norm1''' , '''batchnorm_before''' ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): __lowerCAmelCase : Optional[Any] = new_name.replace('''norm2''' , '''batchnorm_after''' ) if "proj" in new_name: __lowerCAmelCase : str = new_name.replace('''proj''' , '''projection''' ) if "dist_head" in new_name: __lowerCAmelCase : Optional[int] = new_name.replace('''dist_head''' , '''distillation_classifier''' ) elif "head" in new_name: __lowerCAmelCase : Optional[int] = new_name.replace('''head''' , '''classifier''' ) elif "patch_embed" in new_name: __lowerCAmelCase : str = '''efficientformer.''' + new_name elif new_name == "norm.weight" or new_name == "norm.bias": __lowerCAmelCase : Optional[int] = new_name.replace('''norm''' , '''layernorm''' ) __lowerCAmelCase : List[Any] = '''efficientformer.''' + new_name else: __lowerCAmelCase : int = '''efficientformer.encoder.''' + new_name return new_name def _lowercase ( lowercase__ , lowercase__ ): for key in checkpoint.copy().keys(): __lowerCAmelCase : Any = checkpoint.pop(lowercase__ ) __lowerCAmelCase : Tuple = val return checkpoint def _lowercase ( ): __lowerCAmelCase : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowerCAmelCase : Tuple = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return image def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[int] = torch.load(lowercase__ , map_location='''cpu''' )['''model'''] __lowerCAmelCase : Union[str, Any] = EfficientFormerConfig.from_json_file(lowercase__ ) __lowerCAmelCase : Dict = EfficientFormerForImageClassificationWithTeacher(lowercase__ ) __lowerCAmelCase : Optional[Any] = '''_'''.join(checkpoint_path.split('''/''' )[-1].split('''.''' )[0].split('''_''' )[:-1] ) __lowerCAmelCase : int = config.depths[-1] - config.num_metaad_blocks + 1 __lowerCAmelCase : Optional[Any] = convert_torch_checkpoint(lowercase__ , lowercase__ ) model.load_state_dict(lowercase__ ) model.eval() __lowerCAmelCase : int = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } # prepare image __lowerCAmelCase : List[str] = prepare_img() __lowerCAmelCase : int = 2_5_6 __lowerCAmelCase : Optional[Any] = 2_2_4 __lowerCAmelCase : Tuple = EfficientFormerImageProcessor( size={'''shortest_edge''': image_size} , crop_size={'''height''': crop_size, '''width''': crop_size} , resample=pillow_resamplings['''bicubic'''] , ) __lowerCAmelCase : List[str] = processor(images=lowercase__ , return_tensors='''pt''' ).pixel_values # original processing pipeline __lowerCAmelCase : List[str] = Compose( [ Resize(lowercase__ , interpolation=pillow_resamplings['''bicubic'''] ), CenterCrop(lowercase__ ), ToTensor(), Normalize(lowercase__ , lowercase__ ), ] ) __lowerCAmelCase : List[str] = image_transforms(lowercase__ ).unsqueeze(0 ) assert torch.allclose(lowercase__ , lowercase__ ) __lowerCAmelCase : int = model(lowercase__ ) __lowerCAmelCase : Any = outputs.logits __lowerCAmelCase : List[str] = (1, 1_0_0_0) if "l1" in model_name: __lowerCAmelCase : Optional[int] = torch.Tensor( [-0.1_3_1_2, 0.4_3_5_3, -1.0_4_9_9, -0.5_1_2_4, 0.4_1_8_3, -0.6_7_9_3, -1.3_7_7_7, -0.0_8_9_3, -0.7_3_5_8, -2.4_3_2_8] ) assert torch.allclose(logits[0, :1_0] , lowercase__ , atol=1E-3 ) assert logits.shape == expected_shape elif "l3" in model_name: __lowerCAmelCase : str = torch.Tensor( [-1.3_1_5_0, -1.5_4_5_6, -1.2_5_5_6, -0.8_4_9_6, -0.7_1_2_7, -0.7_8_9_7, -0.9_7_2_8, -0.3_0_5_2, 0.3_7_5_1, -0.3_1_2_7] ) assert torch.allclose(logits[0, :1_0] , lowercase__ , atol=1E-3 ) assert logits.shape == expected_shape elif "l7" in model_name: __lowerCAmelCase : Optional[int] = torch.Tensor( [-1.0_2_8_3, -1.4_1_3_1, -0.5_6_4_4, -1.3_1_1_5, -0.5_7_8_5, -1.2_0_4_9, -0.7_5_2_8, 0.1_9_9_2, -0.3_8_2_2, -0.0_8_7_8] ) assert logits.shape == expected_shape else: raise ValueError( f"""Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7""" ) # Save Checkpoints Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) model.save_pretrained(lowercase__ ) print(f"""Checkpoint successfuly converted. Model saved at {pytorch_dump_path}""" ) processor.save_pretrained(lowercase__ ) print(f"""Processor successfuly saved at {pytorch_dump_path}""" ) if push_to_hub: print('''Pushing model to the hub...''' ) model.push_to_hub( repo_id=f"""Bearnardd/{pytorch_dump_path}""" , commit_message='''Add model''' , use_temp_dir=lowercase__ , ) processor.push_to_hub( repo_id=f"""Bearnardd/{pytorch_dump_path}""" , commit_message='''Add image processor''' , use_temp_dir=lowercase__ , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--pytorch_model_path", default=None, type=str, required=True, help="Path to EfficientFormer pytorch checkpoint.", ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The json file for EfficientFormer model config.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Push model and image processor to the hub") parser.add_argument( "--no-push_to_hub", dest="push_to_hub", action="store_false", help="Do not push model and image processor to the hub", ) parser.set_defaults(push_to_hub=True) _UpperCamelCase = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
275
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
1
from math import ceil def _lowercase ( lowercase__ = 1_0_0_1 ): __lowerCAmelCase : Union[str, Any] = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): __lowerCAmelCase : Dict = 2 * i + 1 __lowerCAmelCase : List[str] = 2 * i __lowerCAmelCase : Any = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: _UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print("Invalid entry - please enter a number")
275
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCAmelCase : Dict = Vector() def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(A_ ) , '''(0,0,0,0,0,1)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(A_ ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Vector([1, 2] ) __lowerCAmelCase : Optional[int] = Vector([1, 2, 3, 4, 5] ) __lowerCAmelCase : Optional[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCAmelCase : str = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : List[Any] = Vector([2, -1, 4] ) # for test of dot product __lowerCAmelCase : Optional[int] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : Any = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , A_ , A_ ) ) , '''(3,4,7)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) __lowerCAmelCase : Optional[Any] = x.copy() self.assertEqual(str(A_ ) , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(A_ ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Dict = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
275
1
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase = get_tests_dir("fixtures/spiece.model") @require_sentencepiece @require_tokenizers class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = AlbertTokenizer _UpperCamelCase = AlbertTokenizerFast _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = True def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCAmelCase : Dict = AlbertTokenizer(A_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = '''this is a test''' __lowerCAmelCase : Union[str, Any] = '''this is a test''' return input_text, output_text def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Dict = '''<pad>''' __lowerCAmelCase : Optional[Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ) , A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''<unk>''' ) self.assertEqual(vocab_keys[-1] , '''▁eloquent''' ) self.assertEqual(len(A_ ) , 3_0000 ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 3_0000 ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' if not self.test_rust_tokenizer: return __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : Dict = self.get_rust_tokenizer() __lowerCAmelCase : Tuple = '''I was born in 92000, and this is falsé.''' __lowerCAmelCase : Any = tokenizer.tokenize(A_ ) __lowerCAmelCase : Any = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : int = tokenizer.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : int = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : List[str] = self.get_rust_tokenizer() __lowerCAmelCase : Union[str, Any] = tokenizer.encode(A_ ) __lowerCAmelCase : Any = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Any = AlbertTokenizer(A_ , keep_accents=A_ ) __lowerCAmelCase : Optional[int] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁this''', '''▁is''', '''▁a''', '''▁test'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [48, 25, 21, 1289] ) __lowerCAmelCase : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( A_ , ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''é''', '''.'''] ) __lowerCAmelCase : List[Any] = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual(A_ , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) __lowerCAmelCase : int = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.'''] , ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Dict = AlbertTokenizer(A_ ) __lowerCAmelCase : List[Any] = tokenizer.encode('''sequence builders''' ) __lowerCAmelCase : List[Any] = tokenizer.encode('''multi-sequence build''' ) __lowerCAmelCase : Dict = tokenizer.build_inputs_with_special_tokens(A_ ) __lowerCAmelCase : int = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = {'''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''input_ids''': [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A_ , model_name='''albert-base-v2''' , revision='''6b6560eaf5ff2e250b00c50f380c5389a9c2d82e''' , )
275
def _lowercase ( lowercase__ , lowercase__ ): if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
275
1
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
1
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _UpperCamelCase = logging.getLogger(__name__) _UpperCamelCase = "pytorch_model.bin" @dataclasses.dataclass class __lowercase : _UpperCamelCase = dataclasses.field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models."""} ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co."""} , ) @dataclasses.dataclass class __lowercase : _UpperCamelCase = dataclasses.field(metadata={"""help""": """A csv or a json file containing the training data."""} ) _UpperCamelCase = dataclasses.field(metadata={"""help""": """A csv or a json file containing the data to predict on."""} ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """A csv or a json file containing the validation data."""} ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """The name of the task to train on."""} , ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """The list of labels for the task."""} ) @dataclasses.dataclass class __lowercase : _UpperCamelCase = dataclasses.field( metadata={"""help""": """The output directory where the model predictions and checkpoints will be written."""} ) _UpperCamelCase = dataclasses.field( default="""accuracy""" , metadata={"""help""": """The evaluation metric used for the task."""} ) _UpperCamelCase = dataclasses.field( default="""no""" , metadata={ """help""": """The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]""" } , ) _UpperCamelCase = dataclasses.field( default=10 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) _UpperCamelCase = dataclasses.field( default=0.0 , metadata={ """help""": """How much the specified evaluation metric must improve to satisfy early stopping conditions.""" } , ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the confidence score."""} , ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the validation performance."""} , ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """Whether to fine-tune on labeled data after pseudo training."""} , ) _UpperCamelCase = dataclasses.field( default=0.0 , metadata={"""help""": """Confidence threshold for pseudo-labeled data filtering."""} , ) _UpperCamelCase = dataclasses.field( default=100 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) _UpperCamelCase = dataclasses.field( default=_UpperCAmelCase , metadata={"""help""": """Random seed for initialization."""} , ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: __lowerCAmelCase : Tuple = dataset.filter(lambda lowercase__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 __lowerCAmelCase : Optional[Any] = int(eval_result * len(lowercase__ ) ) print(lowercase__ ) __lowerCAmelCase : int = dataset.sort('''probability''' , reverse=lowercase__ ) __lowerCAmelCase : Union[str, Any] = dataset.select(range(lowercase__ ) ) __lowerCAmelCase : int = dataset.remove_columns(['''label''', '''probability'''] ) __lowerCAmelCase : Any = dataset.rename_column('''prediction''' , '''label''' ) __lowerCAmelCase : Optional[Any] = dataset.map(lambda lowercase__ : {"label": idalabel[example["label"]]} ) __lowerCAmelCase : Any = dataset.shuffle(seed=args.seed ) __lowerCAmelCase : int = os.path.join(lowercase__ , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(lowercase__ , index=lowercase__ ) else: dataset.to_json(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , **lowercase__ ): __lowerCAmelCase : Dict = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() __lowerCAmelCase : List[str] = STModelArguments(model_name_or_path=lowercase__ ) __lowerCAmelCase : List[str] = STDataArguments(train_file=lowercase__ , infer_file=lowercase__ ) __lowerCAmelCase : Any = STTrainingArguments(output_dir=lowercase__ ) __lowerCAmelCase : int = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(lowercase__ ).items(): setattr(lowercase__ , lowercase__ , lowercase__ ) for key, value in kwargs.items(): if hasattr(lowercase__ , lowercase__ ): setattr(lowercase__ , lowercase__ , lowercase__ ) # Sanity checks __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : Any = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None __lowerCAmelCase : int = args.train_file __lowerCAmelCase : str = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None __lowerCAmelCase : str = args.eval_file for key in data_files: __lowerCAmelCase : Optional[int] = data_files[key].split('''.''' )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: __lowerCAmelCase : List[str] = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info('''Creating the initial data directory for self-training...''' ) __lowerCAmelCase : Optional[Any] = f"""{args.output_dir}/self-train_iter-{{}}""".format __lowerCAmelCase : str = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=lowercase__ ) os.makedirs(lowercase__ , exist_ok=lowercase__ ) accelerator.wait_for_everyone() __lowerCAmelCase : int = None __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : Optional[Any] = 0 __lowerCAmelCase : Tuple = False # Show the progress bar __lowerCAmelCase : Dict = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): __lowerCAmelCase : List[str] = data_dir_format(lowercase__ ) assert os.path.exists(lowercase__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 __lowerCAmelCase : int = os.path.join(lowercase__ , '''stage-1''' ) __lowerCAmelCase : Optional[Any] = { '''accelerator''': accelerator, '''model_name_or_path''': args.model_name_or_path, '''cache_dir''': args.cache_dir, '''do_train''': True, '''train_file''': data_files['''train'''] if iteration == 0 else data_files['''train_pseudo'''], '''do_eval''': True if args.eval_file is not None else False, '''eval_file''': data_files['''eval'''], '''do_predict''': True, '''infer_file''': data_files['''infer'''], '''task_name''': args.task_name, '''label_list''': args.label_list, '''output_dir''': current_output_dir, '''eval_metric''': args.eval_metric, '''evaluation_strategy''': args.evaluation_strategy, '''early_stopping_patience''': args.early_stopping_patience, '''early_stopping_threshold''': args.early_stopping_threshold, '''seed''': args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(lowercase__ , lowercase__ ): arguments_dict.update({key: value} ) __lowerCAmelCase : List[Any] = os.path.join(lowercase__ , '''best-checkpoint''' , lowercase__ ) if os.path.exists(lowercase__ ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.''' , lowercase__ , lowercase__ , ) else: logger.info('''***** Running self-training: iteration: %d, stage: 1 *****''' , lowercase__ ) finetune(**lowercase__ ) accelerator.wait_for_everyone() assert os.path.exists(lowercase__ ) logger.info('''Self-training job completed: iteration: %d, stage: 1.''' , lowercase__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data __lowerCAmelCase : str = os.path.join(lowercase__ , '''best-checkpoint''' ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''stage-2''' ) # Update arguments_dict __lowerCAmelCase : int = model_path __lowerCAmelCase : str = data_files['''train'''] __lowerCAmelCase : List[str] = current_output_dir __lowerCAmelCase : Optional[int] = os.path.join(lowercase__ , '''best-checkpoint''' , lowercase__ ) if os.path.exists(lowercase__ ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.''' , lowercase__ , lowercase__ , ) else: logger.info('''***** Running self-training: iteration: %d, stage: 2 *****''' , lowercase__ ) finetune(**lowercase__ ) accelerator.wait_for_everyone() assert os.path.exists(lowercase__ ) logger.info('''Self-training job completed: iteration: %d, stage: 2.''' , lowercase__ ) __lowerCAmelCase : Union[str, Any] = iteration __lowerCAmelCase : Tuple = data_dir_format(iteration + 1 ) __lowerCAmelCase : str = AutoConfig.from_pretrained(os.path.join(lowercase__ , '''best-checkpoint''' ) ) __lowerCAmelCase : List[Any] = config.idalabel __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''eval_results_best-checkpoint.json''' ) __lowerCAmelCase : Tuple = os.path.join(lowercase__ , '''test_results_best-checkpoint.json''' ) assert os.path.exists(lowercase__ ) with open(lowercase__ , '''r''' ) as f: __lowerCAmelCase : List[str] = float(json.load(lowercase__ )[args.eval_metric] ) __lowerCAmelCase : Dict = os.path.join(lowercase__ , '''infer_output_best-checkpoint.csv''' ) assert os.path.exists(lowercase__ ) # Loading the dataset from local csv or json files. __lowerCAmelCase : Any = load_dataset(args.data_file_extension , data_files={'''data''': data_files['''infer''']} )['''data'''] __lowerCAmelCase : Union[str, Any] = load_dataset('''csv''' , data_files={'''data''': infer_output_file} )['''data'''] if accelerator.is_main_process: os.makedirs(lowercase__ , exist_ok=lowercase__ ) shutil.copy(lowercase__ , os.path.join(lowercase__ , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(lowercase__ ): shutil.copy(lowercase__ , os.path.join(lowercase__ , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) accelerator.wait_for_everyone() __lowerCAmelCase : str = os.path.join(lowercase__ , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: __lowerCAmelCase : int = eval_result if best_iteration is None: __lowerCAmelCase : Optional[Any] = new_iteration __lowerCAmelCase : Optional[int] = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: __lowerCAmelCase : str = new_iteration __lowerCAmelCase : str = new_eval_result __lowerCAmelCase : int = 0 else: if new_eval_result == best_eval_result: __lowerCAmelCase : List[str] = new_iteration __lowerCAmelCase : Tuple = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: __lowerCAmelCase : List[str] = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info('''Best iteration: %d''' , lowercase__ ) logger.info('''Best evaluation result: %s = %f''' , args.eval_metric , lowercase__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(lowercase__ , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(lowercase__ , '''eval_results_best-iteration.json''' ) , ) else: # Assume that the last iteration is the best logger.info('''Best iteration: %d''' , args.max_selftrain_iterations - 1 ) logger.info('''Best evaluation result: %s = %f''' , args.eval_metric , lowercase__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(lowercase__ , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(lowercase__ , '''eval_results_best-iteration.json''' ) , )
275
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ConsistencyModelPipeline _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt _UpperCamelCase = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[Any] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def UpperCamelCase__ ( self , A_=False ) ->Dict: '''simple docstring''' if class_cond: __lowerCAmelCase : List[str] = self.dummy_cond_unet else: __lowerCAmelCase : Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->Tuple: '''simple docstring''' if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : str = torch.manual_seed(A_ ) else: __lowerCAmelCase : Dict = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Tuple = self.get_dummy_components() __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : str = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : str = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : int = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Any = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : Union[str, Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : List[str] = 1 __lowerCAmelCase : Dict = None __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , A_=0 , A_=False , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = torch.manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __lowerCAmelCase : List[str] = self.get_fixed_latents(seed=A_ , device=A_ , dtype=A_ , shape=A_ ) __lowerCAmelCase : Union[str, Any] = latents return inputs def UpperCamelCase__ ( self , A_=0 , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->Optional[int]: '''simple docstring''' if type(A_ ) == str: __lowerCAmelCase : int = torch.device(A_ ) __lowerCAmelCase : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Union[str, Any] = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) return latents def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = self.get_inputs() __lowerCAmelCase : Any = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.0_888, 0.0_881, 0.0_666, 0.0_479, 0.0_292, 0.0_195, 0.0_201, 0.0_163, 0.0_254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = self.get_inputs() __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : str = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.0_340, 0.0_152, 0.0_063, 0.0_267, 0.0_221, 0.0_107, 0.0_416, 0.0_186, 0.0_217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_inputs(get_fixed_latents=A_ , device=A_ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.1_875, 0.1_428, 0.1_289, 0.2_151, 0.2_092, 0.1_477, 0.1_877, 0.1_641, 0.1_353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Union[str, Any] = self.get_inputs(get_fixed_latents=A_ , device=A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : int = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : str = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.1_663, 0.1_948, 0.2_275, 0.1_680, 0.1_204, 0.1_245, 0.1_858, 0.1_338, 0.2_095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
275
1
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : str = [] for rt in rc.restypes: __lowerCAmelCase : List[Any] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase : List[str] = {name: i for i, name in enumerate(lowercase__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) __lowerCAmelCase : List[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Optional[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Tuple = torch.tensor( lowercase__ , dtype=torch.floataa , device=protein['''aatype'''].device , ) __lowerCAmelCase : List[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : int = residx_atomaa_mask __lowerCAmelCase : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase : int = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase : str = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase : Optional[int] = rc.restype_atoa[restype_letter] __lowerCAmelCase : Optional[Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase : str = rc.atom_order[atom_name] __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : Any = residx_atomaa_mask return protein def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = tree_map(lambda lowercase__ : torch.tensor(lowercase__ , device=batch['''aatype'''].device ) , lowercase__ , np.ndarray ) __lowerCAmelCase : Tuple = tensor_tree_map(lambda lowercase__ : np.array(lowercase__ ) , make_atomaa_masks(lowercase__ ) ) return out
275
from collections import deque from .hash_table import HashTable class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->int: '''simple docstring''' super().__init__(*A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(A_ ) __lowerCAmelCase : int = self.values[key] def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return ( sum(self.charge_factor - len(A_ ) for slot in self.values ) / self.size_table * self.charge_factor ) def UpperCamelCase__ ( self , A_ , A_=None ) ->str: '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(A_ ) == 0 ): return key return super()._collision_resolution(A_ , A_ )
275
1
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCamelCase = 25_0004 _UpperCamelCase = 25_0020 @require_sentencepiece @require_tokenizers class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = MBartTokenizer _UpperCamelCase = MBartTokenizerFast _UpperCamelCase = True _UpperCamelCase = True def UpperCamelCase__ ( self ) ->str: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCAmelCase : List[Any] = MBartTokenizer(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : str = MBartTokenizer(A_ , keep_accents=A_ ) __lowerCAmelCase : Optional[int] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCAmelCase : str = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCAmelCase : str = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __lowerCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __lowerCAmelCase : Optional[Any] = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Dict = self.tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : List[str] = tempfile.mkdtemp() __lowerCAmelCase : int = tokenizer_r.save_pretrained(A_ ) __lowerCAmelCase : Optional[Any] = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __lowerCAmelCase : Tuple = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way __lowerCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(A_ ) __lowerCAmelCase : Any = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=True __lowerCAmelCase : List[Any] = tempfile.mkdtemp() __lowerCAmelCase : Dict = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) __lowerCAmelCase : List[str] = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way __lowerCAmelCase : Any = tokenizer_r.from_pretrained(A_ ) __lowerCAmelCase : Any = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=False __lowerCAmelCase : Tuple = tempfile.mkdtemp() __lowerCAmelCase : str = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) __lowerCAmelCase : int = tokenizer_p.save_pretrained(A_ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __lowerCAmelCase : str = tokenizer_r.from_pretrained(A_ ) __lowerCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowercase (unittest.TestCase ): _UpperCamelCase = """facebook/mbart-large-en-ro""" _UpperCamelCase = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] _UpperCamelCase = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] _UpperCamelCase = [8274, 127873, 25916, 7, 8622, 2071, 438, 67485, 53, 187895, 23, 51712, 2, EN_CODE] @classmethod def UpperCamelCase__ ( cls ) ->Tuple: '''simple docstring''' __lowerCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) __lowerCAmelCase : Optional[Any] = 1 return cls def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 25_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 25_0004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 25_0020 ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' self.assertIn(A_ , self.tokenizer.all_special_ids ) __lowerCAmelCase : Any = [RO_CODE, 884, 9019, 96, 9, 916, 8_6792, 36, 1_8743, 1_5596, 5, 2] __lowerCAmelCase : Any = self.tokenizer.decode(A_ , skip_special_tokens=A_ ) __lowerCAmelCase : Any = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ ) self.assertNotIn(self.tokenizer.eos_token , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Tuple = ['''this is gunna be a long sentence ''' * 20] assert isinstance(src_text[0] , A_ ) __lowerCAmelCase : int = 10 __lowerCAmelCase : int = self.tokenizer(A_ , max_length=A_ , truncation=A_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , A_ ) self.assertEqual(len(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [25_0026, 25_0001] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = tempfile.mkdtemp() __lowerCAmelCase : List[str] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(A_ ) __lowerCAmelCase : Dict = MBartTokenizer.from_pretrained(A_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , A_ ) @require_torch def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=A_ , return_tensors='''pt''' ) __lowerCAmelCase : Dict = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) __lowerCAmelCase : Union[str, Any] = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(A_ , A_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __lowerCAmelCase : Union[str, Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , A_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.tokenizer(self.src_text , padding=A_ , truncation=A_ , max_length=3 , return_tensors='''pt''' ) __lowerCAmelCase : Optional[Any] = self.tokenizer( text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=10 , return_tensors='''pt''' ) __lowerCAmelCase : Union[str, Any] = targets['''input_ids'''] __lowerCAmelCase : int = shift_tokens_right(A_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(A_ ) , { # A, test, EOS, en_XX '''input_ids''': [[62, 3034, 2, 25_0004]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 25_0001, } , )
275
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
1
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar _UpperCamelCase = TypeVar("T") def _lowercase ( lowercase__ ): return (position - 1) // 2 def _lowercase ( lowercase__ ): return (2 * position) + 1 def _lowercase ( lowercase__ ): return (2 * position) + 2 class __lowercase (Generic[T] ): def __init__( self ) ->None: '''simple docstring''' __lowerCAmelCase : list[tuple[T, int]] = [] __lowerCAmelCase : dict[T, int] = {} __lowerCAmelCase : int = 0 def __len__( self ) ->int: '''simple docstring''' return self.elements def __repr__( self ) ->str: '''simple docstring''' return str(self.heap ) def UpperCamelCase__ ( self ) ->bool: '''simple docstring''' return self.elements == 0 def UpperCamelCase__ ( self , A_ , A_ ) ->None: '''simple docstring''' self.heap.append((elem, weight) ) __lowerCAmelCase : Any = self.elements self.elements += 1 self._bubble_up(A_ ) def UpperCamelCase__ ( self ) ->T: '''simple docstring''' if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: __lowerCAmelCase, __lowerCAmelCase : int = self.heap[0] self._bubble_down(A_ ) return elem def UpperCamelCase__ ( self , A_ , A_ ) ->None: '''simple docstring''' __lowerCAmelCase : str = self.position_map[elem] __lowerCAmelCase : Optional[int] = (elem, weight) if position > 0: __lowerCAmelCase : Optional[Any] = get_parent_position(A_ ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.heap[parent_position] if parent_weight > weight: self._bubble_up(A_ ) else: self._bubble_down(A_ ) else: self._bubble_down(A_ ) def UpperCamelCase__ ( self , A_ ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.position_map[elem] if curr_pos == 0: return None __lowerCAmelCase : Union[str, Any] = get_parent_position(A_ ) __lowerCAmelCase, __lowerCAmelCase : Dict = self.heap[curr_pos] __lowerCAmelCase, __lowerCAmelCase : str = self.heap[parent_position] if parent_weight > weight: self._swap_nodes(A_ , A_ ) return self._bubble_up(A_ ) return None def UpperCamelCase__ ( self , A_ ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = self.position_map[elem] __lowerCAmelCase, __lowerCAmelCase : Tuple = self.heap[curr_pos] __lowerCAmelCase : Union[str, Any] = get_child_left_position(A_ ) __lowerCAmelCase : int = get_child_right_position(A_ ) if child_left_position < self.elements and child_right_position < self.elements: __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.heap[child_left_position] __lowerCAmelCase, __lowerCAmelCase : Optional[int] = self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(A_ , A_ ) return self._bubble_down(A_ ) if child_left_position < self.elements: __lowerCAmelCase, __lowerCAmelCase : Dict = self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(A_ , A_ ) return self._bubble_down(A_ ) else: return None if child_right_position < self.elements: __lowerCAmelCase, __lowerCAmelCase : Tuple = self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(A_ , A_ ) return self._bubble_down(A_ ) return None def UpperCamelCase__ ( self , A_ , A_ ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = self.heap[nodea_pos][0] __lowerCAmelCase : Optional[int] = self.heap[nodea_pos][0] __lowerCAmelCase, __lowerCAmelCase : int = ( self.heap[nodea_pos], self.heap[nodea_pos], ) __lowerCAmelCase : Dict = nodea_pos __lowerCAmelCase : List[str] = nodea_pos class __lowercase (Generic[T] ): def __init__( self ) ->None: '''simple docstring''' __lowerCAmelCase : dict[T, dict[T, int]] = {} __lowerCAmelCase : int = 0 def __repr__( self ) ->str: '''simple docstring''' return str(self.connections ) def __len__( self ) ->int: '''simple docstring''' return self.nodes def UpperCamelCase__ ( self , A_ ) ->None: '''simple docstring''' if node not in self.connections: __lowerCAmelCase : str = {} self.nodes += 1 def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->None: '''simple docstring''' self.add_node(A_ ) self.add_node(A_ ) __lowerCAmelCase : Dict = weight __lowerCAmelCase : str = weight def _lowercase ( lowercase__ , ): __lowerCAmelCase : dict[T, int] = {node: maxsize for node in graph.connections} __lowerCAmelCase : dict[T, T | None] = {node: None for node in graph.connections} __lowerCAmelCase : MinPriorityQueue[T] = MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(lowercase__ , lowercase__ ) if priority_queue.is_empty(): return dist, parent # initialization __lowerCAmelCase : str = priority_queue.extract_min() __lowerCAmelCase : Tuple = 0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __lowerCAmelCase : Tuple = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(lowercase__ , dist[neighbour] ) __lowerCAmelCase : Tuple = node # running prim's algorithm while not priority_queue.is_empty(): __lowerCAmelCase : Any = priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __lowerCAmelCase : Any = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(lowercase__ , dist[neighbour] ) __lowerCAmelCase : Optional[int] = node return dist, parent
275
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=False , A_=True , A_="None" , A_=3 , A_=4 , A_=None , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = parent __lowerCAmelCase : List[str] = batch_size __lowerCAmelCase : Dict = seq_length __lowerCAmelCase : List[Any] = is_training __lowerCAmelCase : List[Any] = use_input_mask __lowerCAmelCase : Optional[int] = use_token_type_ids __lowerCAmelCase : Tuple = use_labels __lowerCAmelCase : str = vocab_size __lowerCAmelCase : int = hidden_size __lowerCAmelCase : Any = num_hidden_layers __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : int = hidden_act __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Union[str, Any] = type_sequence_label_size __lowerCAmelCase : Optional[int] = initializer_range __lowerCAmelCase : int = num_labels __lowerCAmelCase : int = num_choices __lowerCAmelCase : List[str] = relative_attention __lowerCAmelCase : Union[str, Any] = position_biased_input __lowerCAmelCase : int = pos_att_type __lowerCAmelCase : List[Any] = scope def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : int = None if self.use_input_mask: __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __lowerCAmelCase : List[str] = None if self.use_token_type_ids: __lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : int = None __lowerCAmelCase : List[str] = None if self.use_labels: __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_config() __lowerCAmelCase : Dict = 300 return config def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = DebertaModel(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ )[0] __lowerCAmelCase : Any = model(A_ , token_type_ids=A_ )[0] __lowerCAmelCase : List[str] = model(A_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = DebertaForMaskedLM(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = self.num_labels __lowerCAmelCase : Tuple = DebertaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Optional[int] = DebertaForTokenClassification(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Tuple = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = DebertaForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : int = model( A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ) : Tuple = config_and_inputs __lowerCAmelCase : Tuple = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowercase (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : int = DebertaModelTester(self ) __lowerCAmelCase : List[Any] = ConfigTester(self , config_class=A_ , hidden_size=37 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*A_ ) @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[int] = DebertaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowercase (unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' pass @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = DebertaModel.from_pretrained('''microsoft/deberta-base''' ) __lowerCAmelCase : Tuple = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase : Optional[int] = model(A_ , attention_mask=A_ )[0] # compare the actual values for a slice. __lowerCAmelCase : Optional[Any] = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , A_ , atol=1e-4 ) , f"""{output[:, 1:4, 1:4]}""" )
275
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""" ) ) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue_model_parallelism.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1600, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1600, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, ] ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding='''utf-8''' , check=A_ , ) assert hasattr(self , '''env''' ) def UpperCamelCase__ ( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = { '''enabled''': True, '''processes_per_host''': 8, } __lowerCAmelCase : List[Any] = { '''enabled''': True, '''parameters''': { '''microbatches''': 4, '''placement_strategy''': '''spread''', '''pipeline''': '''interleaved''', '''optimize''': '''speed''', '''partitions''': 4, '''ddp''': True, }, } __lowerCAmelCase : int = {'''smdistributed''': {'''modelparallel''': smp_options}, '''mpi''': mpi_options} __lowerCAmelCase : str = '''trainer''' if self.script == '''run_glue.py''' else '''smtrainer''' # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-{instance_count}-smp-{name_extension}""" , instance_count=A_ , instance_type=self.instance_type , debugger_hook_config=A_ , hyperparameters={ **self.env.hyperparameters, '''model_name_or_path''': self.model_name_or_path, '''max_steps''': 500, } , metric_definitions=self.env.metric_definitions , distribution=A_ , py_version='''py36''' , ) def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' TrainingJobAnalytics(A_ ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(1,)] ) def UpperCamelCase__ ( self , A_ ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.create_estimator(A_ ) # run training estimator.fit() # result dataframe __lowerCAmelCase : Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __lowerCAmelCase : List[str] = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Optional[int] = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , '''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , A_ )
275
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : str = [] for rt in rc.restypes: __lowerCAmelCase : List[Any] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase : List[str] = {name: i for i, name in enumerate(lowercase__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) __lowerCAmelCase : List[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Optional[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Tuple = torch.tensor( lowercase__ , dtype=torch.floataa , device=protein['''aatype'''].device , ) __lowerCAmelCase : List[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : int = residx_atomaa_mask __lowerCAmelCase : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase : int = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase : str = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase : Optional[int] = rc.restype_atoa[restype_letter] __lowerCAmelCase : Optional[Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase : str = rc.atom_order[atom_name] __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : Any = residx_atomaa_mask return protein def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = tree_map(lambda lowercase__ : torch.tensor(lowercase__ , device=batch['''aatype'''].device ) , lowercase__ , np.ndarray ) __lowerCAmelCase : Tuple = tensor_tree_map(lambda lowercase__ : np.array(lowercase__ ) , make_atomaa_masks(lowercase__ ) ) return out
275
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _UpperCamelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["MLukeTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
def _lowercase ( lowercase__ ): if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) __lowerCAmelCase : int = sorted(string.lower() ) return len(lowercase__ ) == len(set(lowercase__ ) ) if __name__ == "__main__": _UpperCamelCase = input("Enter a string ").strip() _UpperCamelCase = is_isogram(input_str) print(F"{input_str} is {'an' if isogram else 'not an'} isogram.")
275
1
from __future__ import annotations def _lowercase ( lowercase__ , lowercase__ ): if len(lowercase__ ) < k or k < 0: raise ValueError('''Invalid Input''' ) __lowerCAmelCase : Tuple = sum(array[:k] ) for i in range(len(lowercase__ ) - k ): __lowerCAmelCase : int = current_sum - array[i] + array[i + k] __lowerCAmelCase : List[Any] = max(lowercase__ , lowercase__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() _UpperCamelCase = [randint(-1000, 1000) for i in range(100)] _UpperCamelCase = randint(0, 110) print(F"The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}")
275
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = ('''dense.weight''', '''attention.self.query''', '''attention.self.key''', '''attention.self.value''') __lowerCAmelCase : Any = ( ('''layer.''', '''layer_'''), ('''word_embeddings.weight''', '''word_embeddings'''), ('''position_embeddings.weight''', '''position_embeddings'''), ('''token_type_embeddings.weight''', '''token_type_embeddings'''), ('''.''', '''/'''), ('''LayerNorm/weight''', '''LayerNorm/gamma'''), ('''LayerNorm/bias''', '''LayerNorm/beta'''), ('''weight''', '''kernel'''), ) if not os.path.isdir(lowercase__ ): os.makedirs(lowercase__ ) __lowerCAmelCase : Any = model.state_dict() def to_tf_var_name(lowercase__ ): for patt, repl in iter(lowercase__ ): __lowerCAmelCase : Union[str, Any] = name.replace(lowercase__ , lowercase__ ) return f"""bert/{name}""" def create_tf_var(lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = tf.dtypes.as_dtype(tensor.dtype ) __lowerCAmelCase : List[Any] = tf.get_variable(dtype=lowercase__ , shape=tensor.shape , name=lowercase__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(lowercase__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __lowerCAmelCase : Tuple = to_tf_var_name(lowercase__ ) __lowerCAmelCase : List[Any] = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): __lowerCAmelCase : str = torch_tensor.T __lowerCAmelCase : Optional[int] = create_tf_var(tensor=lowercase__ , name=lowercase__ , session=lowercase__ ) tf.keras.backend.set_value(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = session.run(lowercase__ ) print(f"""Successfully created {tf_name}: {np.allclose(lowercase__ , lowercase__ )}""" ) __lowerCAmelCase : Any = tf.train.Saver(tf.trainable_variables() ) saver.save(lowercase__ , os.path.join(lowercase__ , model_name.replace('''-''' , '''_''' ) + '''.ckpt''' ) ) def _lowercase ( lowercase__=None ): __lowerCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument('''--model_name''' , type=lowercase__ , required=lowercase__ , help='''model name e.g. bert-base-uncased''' ) parser.add_argument( '''--cache_dir''' , type=lowercase__ , default=lowercase__ , required=lowercase__ , help='''Directory containing pytorch model''' ) parser.add_argument('''--pytorch_model_path''' , type=lowercase__ , required=lowercase__ , help='''/path/to/<pytorch-model-name>.bin''' ) parser.add_argument('''--tf_cache_dir''' , type=lowercase__ , required=lowercase__ , help='''Directory in which to save tensorflow model''' ) __lowerCAmelCase : Dict = parser.parse_args(lowercase__ ) __lowerCAmelCase : Optional[int] = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=lowercase__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
275
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = np.full((len(lowercase__ ), sequence_length, 2) , lowercase__ ) else: __lowerCAmelCase : Optional[int] = np.full((len(lowercase__ ), sequence_length) , lowercase__ ) for i, tensor in enumerate(lowercase__ ): if padding_side == "right": if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : int = tensor[:sequence_length] else: if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = ord(lowercase__ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True __lowerCAmelCase : int = unicodedata.category(lowercase__ ) if cat.startswith('''P''' ): return True return False @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = True _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = -100 _UpperCamelCase = "pt" def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCAmelCase : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCAmelCase : List[Any] = self.tokenizer.pad( A_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCAmelCase : Dict = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCAmelCase : Optional[int] = self.tokenizer.padding_side if padding_side == "right": __lowerCAmelCase : Any = [ list(A_ ) + [self.label_pad_token_id] * (sequence_length - len(A_ )) for label in labels ] else: __lowerCAmelCase : Optional[int] = [ [self.label_pad_token_id] * (sequence_length - len(A_ )) + list(A_ ) for label in labels ] __lowerCAmelCase : Tuple = [feature['''ner_tags'''] for feature in features] __lowerCAmelCase : List[Any] = padding_tensor(A_ , -1 , A_ , A_ ) __lowerCAmelCase : Optional[int] = [feature['''original_entity_spans'''] for feature in features] __lowerCAmelCase : Any = padding_tensor(A_ , (-1, -1) , A_ , A_ ) __lowerCAmelCase : Optional[Any] = {k: torch.tensor(A_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
275
1
def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Union[str, Any] = len(lowercase__ ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCAmelCase : Dict = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowercase__ ): return None __lowerCAmelCase : int = sorted_collection[point] if current_item == item: return point else: if point < left: __lowerCAmelCase : Dict = left __lowerCAmelCase : int = point elif point > right: __lowerCAmelCase : Any = right __lowerCAmelCase : Optional[Any] = point else: if item < current_item: __lowerCAmelCase : Any = point - 1 else: __lowerCAmelCase : List[Any] = point + 1 return None def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __lowerCAmelCase : Optional[int] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowercase__ ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) elif point > right: return interpolation_search_by_recursion(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( lowercase__ , lowercase__ , lowercase__ , point - 1 ) else: return interpolation_search_by_recursion( lowercase__ , lowercase__ , point + 1 , lowercase__ ) def _lowercase ( lowercase__ ): if collection != sorted(lowercase__ ): raise ValueError('''Collection must be ascending sorted''' ) return True if __name__ == "__main__": import sys _UpperCamelCase = 0 if debug == 1: _UpperCamelCase = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") _UpperCamelCase = 67 _UpperCamelCase = interpolation_search(collection, target) if result is not None: print(F"{target} found at positions: {result}") else: print("Not found")
275
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase (unittest.TestCase ): _UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Tuple = (3, 32, 128) __lowerCAmelCase : List[str] = tempfile.mkdtemp() # fmt: off __lowerCAmelCase : List[str] = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on __lowerCAmelCase : Optional[int] = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) __lowerCAmelCase : Union[str, Any] = { '''do_normalize''': False, '''do_resize''': True, '''image_processor_type''': '''ViTImageProcessor''', '''resample''': 3, '''size''': {'''height''': 32, '''width''': 128}, } __lowerCAmelCase : Optional[Any] = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(A_ , A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) __lowerCAmelCase : str = Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) return image_input def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Union[str, Any] = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : int = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) __lowerCAmelCase : int = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.get_image_processor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Optional[int] = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = image_processor(A_ , return_tensors='''np''' ) __lowerCAmelCase : Tuple = processor(images=A_ , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Union[str, Any] = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Any = '''test''' __lowerCAmelCase : Dict = processor(text=A_ ) __lowerCAmelCase : str = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : str = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = '''test''' __lowerCAmelCase : int = self.prepare_image_inputs() __lowerCAmelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''pixel_values''', '''labels'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Optional[int] = processor.char_decode(A_ ) __lowerCAmelCase : Tuple = tokenizer.batch_decode(A_ ) __lowerCAmelCase : Any = [seq.replace(''' ''' , '''''' ) for seq in decoded_tok] self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = self.prepare_image_inputs() __lowerCAmelCase : List[Any] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_image_processor() __lowerCAmelCase : List[str] = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = torch.randn(1 , 27 , 38 ) __lowerCAmelCase : Optional[int] = torch.randn(1 , 27 , 5_0257 ) __lowerCAmelCase : Optional[Any] = torch.randn(1 , 27 , 3_0522 ) __lowerCAmelCase : List[str] = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['''generated_text''', '''scores''', '''char_preds''', '''bpe_preds''', '''wp_preds'''] )
275
1
from __future__ import annotations import numpy as np def _lowercase ( lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : Optional[int] = np.shape(lowercase__ ) if rows != columns: __lowerCAmelCase : Tuple = ( '''\'table\' has to be of square shaped array but got a ''' f"""{rows}x{columns} array:\n{table}""" ) raise ValueError(lowercase__ ) __lowerCAmelCase : Optional[Any] = np.zeros((rows, columns) ) __lowerCAmelCase : Union[str, Any] = np.zeros((rows, columns) ) for i in range(lowercase__ ): for j in range(lowercase__ ): __lowerCAmelCase : Union[str, Any] = sum(lower[i][k] * upper[k][j] for k in range(lowercase__ ) ) if upper[j][j] == 0: raise ArithmeticError('''No LU decomposition exists''' ) __lowerCAmelCase : Union[str, Any] = (table[i][j] - total) / upper[j][j] __lowerCAmelCase : Dict = 1 for j in range(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = sum(lower[i][k] * upper[k][j] for k in range(lowercase__ ) ) __lowerCAmelCase : Any = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
275
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = self.dummy_uncond_unet __lowerCAmelCase : Any = PNDMScheduler() __lowerCAmelCase : Dict = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' ).images __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''google/ddpm-cifar10-32''' __lowerCAmelCase : Union[str, Any] = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : int = PNDMScheduler() __lowerCAmelCase : Any = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
1
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _UpperCamelCase = abspath(join(dirname(__file__), "src")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="ignore", category=FutureWarning) def _lowercase ( lowercase__ ): config.addinivalue_line( '''markers''' , '''is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested''' ) config.addinivalue_line( '''markers''' , '''is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested''' ) config.addinivalue_line('''markers''' , '''is_pipeline_test: mark test to run only when pipelines are tested''' ) config.addinivalue_line('''markers''' , '''is_staging_test: mark test to run only in the staging environment''' ) config.addinivalue_line('''markers''' , '''accelerate_tests: mark test that require accelerate''' ) config.addinivalue_line('''markers''' , '''tool_tests: mark the tool tests that are run on their specific schedule''' ) def _lowercase ( lowercase__ ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase__ ) def _lowercase ( lowercase__ ): from transformers.testing_utils import pytest_terminal_summary_main __lowerCAmelCase : int = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase__ , id=lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: __lowerCAmelCase : Optional[Any] = 0 # Doctest custom flag to ignore output. _UpperCamelCase = doctest.register_optionflag("IGNORE_RESULT") _UpperCamelCase = doctest.OutputChecker class __lowercase (_UpperCAmelCase ): def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Union[str, Any]: '''simple docstring''' if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , A_ , A_ , A_ ) _UpperCamelCase = CustomOutputChecker _UpperCamelCase = HfDoctestModule _UpperCamelCase = HfDocTestParser
275
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
1
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class __lowercase (pl.LightningModule ): def __init__( self , A_ ) ->List[Any]: '''simple docstring''' super().__init__() __lowerCAmelCase : List[str] = model __lowerCAmelCase : int = 2 __lowerCAmelCase : Optional[Any] = nn.Linear(self.model.config.hidden_size , self.num_labels ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' pass def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # load longformer model from model identifier __lowerCAmelCase : Any = LongformerModel.from_pretrained(lowercase__ ) __lowerCAmelCase : List[str] = LightningModel(lowercase__ ) __lowerCAmelCase : Any = torch.load(lowercase__ , map_location=torch.device('''cpu''' ) ) lightning_model.load_state_dict(ckpt['''state_dict'''] ) # init longformer question answering model __lowerCAmelCase : Any = LongformerForQuestionAnswering.from_pretrained(lowercase__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(lowercase__ ) print(f"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--longformer_model", default=None, type=str, required=True, help="model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.", ) parser.add_argument( "--longformer_question_answering_ckpt_path", default=None, type=str, required=True, help="Path the official PyTorch Lightning Checkpoint.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _UpperCamelCase = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
275
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _UpperCamelCase = { "AI-Sweden/gpt-sw3-126m": 2048, "AI-Sweden/gpt-sw3-350m": 2048, "AI-Sweden/gpt-sw3-1.6b": 2048, "AI-Sweden/gpt-sw3-6.7b": 2048, "AI-Sweden/gpt-sw3-20b": 2048, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_ , A_=False , A_=False , A_=False , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCAmelCase : int = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCAmelCase : Union[str, Any] = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCAmelCase : str = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCAmelCase : Any = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token __lowerCAmelCase : int = eos_token if bos_token is None else bos_token else: __lowerCAmelCase : Optional[int] = '''<pad>''' if pad_token is None else pad_token __lowerCAmelCase : List[str] = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=A_ , remove_space=A_ , keep_accents=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Union[str, Any] = do_lower_case __lowerCAmelCase : Union[str, Any] = remove_space __lowerCAmelCase : int = keep_accents __lowerCAmelCase : Union[str, Any] = vocab_file __lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # Used for whitespace normalization in input texts # fmt : off __lowerCAmelCase : List[Any] = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCAmelCase : int = re.compile( f"""[{"".join(map(A_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.__dict__.copy() __lowerCAmelCase : List[Any] = None return state def __setstate__( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.non_printing_characters_re.sub('''''' , A_ ) # Normalize whitespaces __lowerCAmelCase : List[str] = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCAmelCase : Tuple = unicodedata.normalize('''NFC''' , A_ ) return text def UpperCamelCase__ ( self , A_ , **A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = self.preprocess_text(A_ ) return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' return self.sp_model.PieceToId(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.IdToPiece(A_ ) @staticmethod def UpperCamelCase__ ( A_ ) ->str: '''simple docstring''' return out_string def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = [] __lowerCAmelCase : Tuple = '''''' __lowerCAmelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string def UpperCamelCase__ ( self ) ->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : str = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Any = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = False ) ->Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[Any] = self.preprocess_text(A_ ) __lowerCAmelCase : Dict = self.sp_model.encode(A_ ) else: __lowerCAmelCase : Dict = [self.preprocess_text(A_ ) for t in text] __lowerCAmelCase : Optional[int] = self.sp_model.encode(A_ ) if return_tensors is True or return_tensors == "pt": __lowerCAmelCase : Tuple = torch.tensor(A_ ) return token_ids def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.decode(A_ ) def UpperCamelCase__ ( self , A_ ) ->List[int]: '''simple docstring''' __lowerCAmelCase : int = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCAmelCase : Any = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(A_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=A_ )
275
1
from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging _UpperCamelCase = logging.get_logger(__name__) class __lowercase : _UpperCamelCase = 42 _UpperCamelCase = None @staticmethod def UpperCamelCase__ ( ) ->int: '''simple docstring''' raise NotImplementedError def UpperCamelCase__ ( self , A_ , A_ , A_ , **A_ ) ->List[str]: '''simple docstring''' raise NotImplementedError def UpperCamelCase__ ( self , A_ ) ->List[str]: '''simple docstring''' raise NotImplementedError def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' if not self.is_available(): raise RuntimeError( f"""You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}.""" ) @classmethod def UpperCamelCase__ ( cls ) ->int: '''simple docstring''' return f"""`pip install {cls.pip_package or cls.name}`""" class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """optuna""" @staticmethod def UpperCamelCase__ ( ) ->List[str]: '''simple docstring''' return is_optuna_available() def UpperCamelCase__ ( self , A_ , A_ , A_ , **A_ ) ->Any: '''simple docstring''' return run_hp_search_optuna(A_ , A_ , A_ , **A_ ) def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' return default_hp_space_optuna(A_ ) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """ray""" _UpperCamelCase = """'ray[tune]'""" @staticmethod def UpperCamelCase__ ( ) ->List[Any]: '''simple docstring''' return is_ray_available() def UpperCamelCase__ ( self , A_ , A_ , A_ , **A_ ) ->Dict: '''simple docstring''' return run_hp_search_ray(A_ , A_ , A_ , **A_ ) def UpperCamelCase__ ( self , A_ ) ->Dict: '''simple docstring''' return default_hp_space_ray(A_ ) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """sigopt""" @staticmethod def UpperCamelCase__ ( ) ->Dict: '''simple docstring''' return is_sigopt_available() def UpperCamelCase__ ( self , A_ , A_ , A_ , **A_ ) ->List[Any]: '''simple docstring''' return run_hp_search_sigopt(A_ , A_ , A_ , **A_ ) def UpperCamelCase__ ( self , A_ ) ->Dict: '''simple docstring''' return default_hp_space_sigopt(A_ ) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """wandb""" @staticmethod def UpperCamelCase__ ( ) ->Any: '''simple docstring''' return is_wandb_available() def UpperCamelCase__ ( self , A_ , A_ , A_ , **A_ ) ->Union[str, Any]: '''simple docstring''' return run_hp_search_wandb(A_ , A_ , A_ , **A_ ) def UpperCamelCase__ ( self , A_ ) ->List[str]: '''simple docstring''' return default_hp_space_wandb(A_ ) _UpperCamelCase = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def _lowercase ( ): __lowerCAmelCase : List[str] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(lowercase__ ) > 0: __lowerCAmelCase : Tuple = available_backends[0].name if len(lowercase__ ) > 1: logger.info( f"""{len(lowercase__ )} hyperparameter search backends available. Using {name} as the default.""" ) return name raise RuntimeError( '''No hyperparameter search backend available.\n''' + '''\n'''.join( f""" - To install {backend.name} run {backend.pip_install()}""" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
275
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """table-transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) ->Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowerCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(A_ , A_ ): __lowerCAmelCase : int = backbone_config.get('''model_type''' ) __lowerCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __lowerCAmelCase : Any = config_class.from_dict(A_ ) # set timm attributes to None __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = None, None, None __lowerCAmelCase : Tuple = use_timm_backbone __lowerCAmelCase : Optional[Any] = backbone_config __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = num_queries __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_ffn_dim __lowerCAmelCase : Optional[int] = encoder_layers __lowerCAmelCase : List[str] = encoder_attention_heads __lowerCAmelCase : str = decoder_ffn_dim __lowerCAmelCase : Union[str, Any] = decoder_layers __lowerCAmelCase : Any = decoder_attention_heads __lowerCAmelCase : Optional[int] = dropout __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : Tuple = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : List[str] = init_std __lowerCAmelCase : Tuple = init_xavier_std __lowerCAmelCase : Any = encoder_layerdrop __lowerCAmelCase : List[Any] = decoder_layerdrop __lowerCAmelCase : Optional[Any] = encoder_layers __lowerCAmelCase : Optional[Any] = auxiliary_loss __lowerCAmelCase : Optional[Any] = position_embedding_type __lowerCAmelCase : Tuple = backbone __lowerCAmelCase : Any = use_pretrained_backbone __lowerCAmelCase : int = dilation # Hungarian matcher __lowerCAmelCase : Dict = class_cost __lowerCAmelCase : List[str] = bbox_cost __lowerCAmelCase : int = giou_cost # Loss coefficients __lowerCAmelCase : Optional[Any] = mask_loss_coefficient __lowerCAmelCase : Tuple = dice_loss_coefficient __lowerCAmelCase : int = bbox_loss_coefficient __lowerCAmelCase : List[Any] = giou_loss_coefficient __lowerCAmelCase : int = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.d_model class __lowercase (_UpperCAmelCase ): _UpperCamelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def UpperCamelCase__ ( self ) ->float: '''simple docstring''' return 1e-5 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 12
275
1
_UpperCamelCase = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" _UpperCamelCase = [{"type": "code", "content": INSTALL_CONTENT}] _UpperCamelCase = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=1 / 255 , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1333} __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Optional[int] = num_channels __lowerCAmelCase : List[str] = min_resolution __lowerCAmelCase : List[Any] = max_resolution __lowerCAmelCase : str = do_resize __lowerCAmelCase : Dict = size __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : List[str] = rescale_factor __lowerCAmelCase : Tuple = do_normalize __lowerCAmelCase : str = image_mean __lowerCAmelCase : Union[str, Any] = image_std __lowerCAmelCase : Optional[Any] = do_pad def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def UpperCamelCase__ ( self , A_ , A_=False ) ->List[str]: '''simple docstring''' if not batched: __lowerCAmelCase : str = image_inputs[0] if isinstance(A_ , Image.Image ): __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = image.size else: __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase : Optional[Any] = int(self.size['''shortest_edge'''] * h / w ) __lowerCAmelCase : Union[str, Any] = self.size['''shortest_edge'''] elif w > h: __lowerCAmelCase : Optional[Any] = self.size['''shortest_edge'''] __lowerCAmelCase : List[Any] = int(self.size['''shortest_edge'''] * w / h ) else: __lowerCAmelCase : Union[str, Any] = self.size['''shortest_edge'''] __lowerCAmelCase : int = self.size['''shortest_edge'''] else: __lowerCAmelCase : Dict = [] for image in image_inputs: __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase : Tuple = max(A_ , key=lambda A_ : item[0] )[0] __lowerCAmelCase : List[str] = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = DetrImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = DetrImageProcessingTester(self ) @property def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_rescale''' ) ) self.assertTrue(hasattr(A_ , '''rescale_factor''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) self.assertTrue(hasattr(A_ , '''do_pad''' ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) __lowerCAmelCase : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' pass def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input __lowerCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCAmelCase, __lowerCAmelCase : Dict = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase, __lowerCAmelCase : Tuple = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) __lowerCAmelCase : Dict = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input __lowerCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCAmelCase, __lowerCAmelCase : str = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : Optional[int] = image_processing(A_ , return_tensors='''pt''' ).pixel_values __lowerCAmelCase, __lowerCAmelCase : List[str] = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input __lowerCAmelCase : int = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCAmelCase, __lowerCAmelCase : Dict = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : List[str] = image_processing(A_ , return_tensors='''pt''' ).pixel_values __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: __lowerCAmelCase : str = json.loads(f.read() ) __lowerCAmelCase : List[str] = {'''image_id''': 3_9769, '''annotations''': target} # encode them __lowerCAmelCase : Dict = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50''' ) __lowerCAmelCase : str = image_processing(images=A_ , annotations=A_ , return_tensors='''pt''' ) # verify pixel values __lowerCAmelCase : Dict = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) __lowerCAmelCase : Union[str, Any] = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1e-4 ) ) # verify area __lowerCAmelCase : Optional[int] = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes __lowerCAmelCase : Tuple = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) __lowerCAmelCase : List[str] = torch.tensor([0.5_503, 0.2_765, 0.0_604, 0.2_215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1e-3 ) ) # verify image_id __lowerCAmelCase : int = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd __lowerCAmelCase : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels __lowerCAmelCase : List[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify orig_size __lowerCAmelCase : Union[str, Any] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size __lowerCAmelCase : List[str] = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) ) @slow def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: __lowerCAmelCase : str = json.loads(f.read() ) __lowerCAmelCase : str = {'''file_name''': '''000000039769.png''', '''image_id''': 3_9769, '''segments_info''': target} __lowerCAmelCase : Optional[int] = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them __lowerCAmelCase : List[Any] = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50-panoptic''' ) __lowerCAmelCase : Optional[Any] = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='''pt''' ) # verify pixel values __lowerCAmelCase : int = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , A_ ) __lowerCAmelCase : List[Any] = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A_ , atol=1e-4 ) ) # verify area __lowerCAmelCase : int = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A_ ) ) # verify boxes __lowerCAmelCase : str = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A_ ) __lowerCAmelCase : str = torch.tensor([0.2_625, 0.5_437, 0.4_688, 0.8_625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A_ , atol=1e-3 ) ) # verify image_id __lowerCAmelCase : Dict = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A_ ) ) # verify is_crowd __lowerCAmelCase : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A_ ) ) # verify class_labels __lowerCAmelCase : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A_ ) ) # verify masks __lowerCAmelCase : int = 82_2873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A_ ) # verify orig_size __lowerCAmelCase : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A_ ) ) # verify size __lowerCAmelCase : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A_ ) )
275
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = {} with open(lowercase__ , '''r''' ) as file: for line_number, line in enumerate(lowercase__ ): __lowerCAmelCase : Any = line.strip() if line: __lowerCAmelCase : Dict = line.split() __lowerCAmelCase : str = line_number __lowerCAmelCase : List[str] = words[0] __lowerCAmelCase : Any = value return result def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[Any] = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : Tuple = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : List[Any] = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : str = getattr(lowercase__ , lowercase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCAmelCase : Dict = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : str = shape_pointer.shape # let's reduce dimension __lowerCAmelCase : Any = value[0] else: __lowerCAmelCase : str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __lowerCAmelCase : List[str] = value elif weight_type == "weight_v": __lowerCAmelCase : int = value elif weight_type == "bias": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : int = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : Tuple = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCAmelCase : List[str] = '''.'''.join([key, hf_param_name] ) else: __lowerCAmelCase : Optional[int] = key __lowerCAmelCase : Union[str, Any] = value if '''lm_head''' in full_key else value[0] _UpperCamelCase = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None ): __lowerCAmelCase : Any = False for key, mapped_key in MAPPING.items(): __lowerCAmelCase : Tuple = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : List[str] = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : List[Any] = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : Any = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCAmelCase : int = '''weight''' else: __lowerCAmelCase : Any = None if hf_dict is not None: rename_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return is_used return is_used def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[Any] = fairseq_model.state_dict() __lowerCAmelCase : Tuple = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : int = True else: __lowerCAmelCase : Dict = load_wavaveca_layer(lowercase__ , lowercase__ , lowercase__ ) if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : List[str] = name.split('''.''' ) __lowerCAmelCase : Any = int(items[0] ) __lowerCAmelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=True , lowercase__=False ): if config_path is not None: __lowerCAmelCase : Union[str, Any] = WavaVecaConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Optional[int] = WavaVecaConfig() if is_seq_class: __lowerCAmelCase : Optional[Any] = read_txt_into_dict(lowercase__ ) __lowerCAmelCase : int = idalabel __lowerCAmelCase : Optional[int] = WavaVecaForSequenceClassification(lowercase__ ) __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) feature_extractor.save_pretrained(lowercase__ ) elif is_finetuned: if dict_path: __lowerCAmelCase : List[str] = Dictionary.load(lowercase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCAmelCase : List[Any] = target_dict.pad_index __lowerCAmelCase : List[Any] = target_dict.bos_index __lowerCAmelCase : Optional[int] = target_dict.eos_index __lowerCAmelCase : Any = len(target_dict.symbols ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''vocab.json''' ) if not os.path.isdir(lowercase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(lowercase__ ) ) return os.makedirs(lowercase__ , exist_ok=lowercase__ ) __lowerCAmelCase : Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : int = 1 with open(lowercase__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = WavaVecaCTCTokenizer( lowercase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=lowercase__ , ) __lowerCAmelCase : List[str] = True if config.feat_extract_norm == '''layer''' else False __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=lowercase__ , tokenizer=lowercase__ ) processor.save_pretrained(lowercase__ ) __lowerCAmelCase : str = WavaVecaForCTC(lowercase__ ) else: __lowerCAmelCase : Any = WavaVecaForPreTraining(lowercase__ ) if is_finetuned or is_seq_class: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCAmelCase : Union[str, Any] = argparse.Namespace(task='''audio_pretraining''' ) __lowerCAmelCase : str = fairseq.tasks.setup_task(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowercase__ ) __lowerCAmelCase : int = model[0].eval() recursively_load_weights(lowercase__ , lowercase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
275
1
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''laion/clap-htsat-unfused''' __lowerCAmelCase : int = tempfile.mkdtemp() def UpperCamelCase__ ( self , **A_ ) ->Optional[Any]: '''simple docstring''' return RobertaTokenizer.from_pretrained(self.checkpoint , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->List[str]: '''simple docstring''' return ClapFeatureExtractor.from_pretrained(self.checkpoint , **A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : Optional[Any] = ClapProcessor(tokenizer=A_ , feature_extractor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Any = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : str = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : Any = self.get_feature_extractor(do_normalize=A_ , padding_value=1.0 ) __lowerCAmelCase : Dict = ClapProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.get_feature_extractor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Dict = ClapProcessor(tokenizer=A_ , feature_extractor=A_ ) __lowerCAmelCase : str = floats_list((3, 1000) ) __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ) __lowerCAmelCase : int = processor(audios=A_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = self.get_feature_extractor() __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : Any = ClapProcessor(tokenizer=A_ , feature_extractor=A_ ) __lowerCAmelCase : Any = '''This is a test string''' __lowerCAmelCase : Optional[int] = processor(text=A_ ) __lowerCAmelCase : Tuple = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.get_feature_extractor() __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = ClapProcessor(tokenizer=A_ , feature_extractor=A_ ) __lowerCAmelCase : Tuple = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Any = processor.batch_decode(A_ ) __lowerCAmelCase : Any = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.get_feature_extractor() __lowerCAmelCase : Union[str, Any] = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = ClapProcessor(tokenizer=A_ , feature_extractor=A_ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , )
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = UnCLIPImageVariationPipeline _UpperCamelCase = IMAGE_VARIATION_PARAMS - {"""height""", """width""", """guidance_scale"""} _UpperCamelCase = IMAGE_VARIATION_BATCH_PARAMS _UpperCamelCase = [ """generator""", """return_dict""", """decoder_num_inference_steps""", """super_res_num_inference_steps""", ] _UpperCamelCase = False @property def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' return 32 @property def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' return 32 @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return self.time_input_dim @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return 100 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(A_ ) @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : Optional[int] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(A_ ) @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : Optional[Any] = { '''clip_embeddings_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''cross_attention_dim''': self.cross_attention_dim, } __lowerCAmelCase : Any = UnCLIPTextProjModel(**A_ ) return model @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : int = { '''sample_size''': 32, # RGB in channels '''in_channels''': 3, # Out channels is double in channels because predicts mean and variance '''out_channels''': 6, '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': '''identity''', } __lowerCAmelCase : Any = UNetaDConditionModel(**A_ ) return model @property def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : str = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' torch.manual_seed(1 ) __lowerCAmelCase : List[Any] = UNetaDModel(**self.dummy_super_res_kwargs ) return model def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.dummy_decoder __lowerCAmelCase : Optional[Any] = self.dummy_text_proj __lowerCAmelCase : Dict = self.dummy_text_encoder __lowerCAmelCase : Union[str, Any] = self.dummy_tokenizer __lowerCAmelCase : Optional[int] = self.dummy_super_res_first __lowerCAmelCase : List[Any] = self.dummy_super_res_last __lowerCAmelCase : List[Any] = UnCLIPScheduler( variance_type='''learned_range''' , prediction_type='''epsilon''' , num_train_timesteps=1000 , ) __lowerCAmelCase : Optional[int] = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''epsilon''' , num_train_timesteps=1000 , ) __lowerCAmelCase : Optional[Any] = CLIPImageProcessor(crop_size=32 , size=32 ) __lowerCAmelCase : Tuple = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def UpperCamelCase__ ( self , A_ , A_=0 , A_=True ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[str] = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_ ) ).to(A_ ) if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : str = torch.manual_seed(A_ ) else: __lowerCAmelCase : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) if pil_image: __lowerCAmelCase : Union[str, Any] = input_image * 0.5 + 0.5 __lowerCAmelCase : Dict = input_image.clamp(0 , 1 ) __lowerCAmelCase : Any = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __lowerCAmelCase : List[Any] = DiffusionPipeline.numpy_to_pil(A_ )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = '''cpu''' __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = self.pipeline_class(**A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[str] = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : Dict = pipe(**A_ ) __lowerCAmelCase : List[Any] = output.images __lowerCAmelCase : int = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : List[str] = pipe( **A_ , return_dict=A_ , )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[Any] = np.array( [ 0.9_997, 0.0_002, 0.9_997, 0.9_997, 0.9_969, 0.0_023, 0.9_997, 0.9_969, 0.9_970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = '''cpu''' __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : List[str] = self.pipeline_class(**A_ ) __lowerCAmelCase : Optional[int] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[int] = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : Union[str, Any] = pipe(**A_ ) __lowerCAmelCase : str = output.images __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : Optional[Any] = pipe( **A_ , return_dict=A_ , )[0] __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : str = np.array([0.9_997, 0.0_003, 0.9_997, 0.9_997, 0.9_970, 0.0_024, 0.9_997, 0.9_971, 0.9_971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''cpu''' __lowerCAmelCase : int = self.get_dummy_components() __lowerCAmelCase : int = self.pipeline_class(**A_ ) __lowerCAmelCase : int = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : Optional[int] = [ pipeline_inputs['''image'''], pipeline_inputs['''image'''], ] __lowerCAmelCase : Dict = pipe(**A_ ) __lowerCAmelCase : Dict = output.images __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : int = [ tuple_pipeline_inputs['''image'''], tuple_pipeline_inputs['''image'''], ] __lowerCAmelCase : Dict = pipe( **A_ , return_dict=A_ , )[0] __lowerCAmelCase : List[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) __lowerCAmelCase : Any = np.array( [ 0.9_997, 0.9_989, 0.0_008, 0.0_021, 0.9_960, 0.0_018, 0.0_014, 0.0_002, 0.9_933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = torch.device('''cpu''' ) class __lowercase : _UpperCamelCase = 1 __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : Optional[int] = self.pipeline_class(**A_ ) __lowerCAmelCase : Union[str, Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = torch.Generator(device=A_ ).manual_seed(0 ) __lowerCAmelCase : List[str] = pipe.decoder.dtype __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Optional[Any] = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) __lowerCAmelCase : List[str] = pipe.prepare_latents( A_ , dtype=A_ , device=A_ , generator=A_ , latents=A_ , scheduler=DummyScheduler() ) __lowerCAmelCase : Optional[int] = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) __lowerCAmelCase : int = pipe.prepare_latents( A_ , dtype=A_ , device=A_ , generator=A_ , latents=A_ , scheduler=DummyScheduler() ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(A_ , pil_image=A_ ) __lowerCAmelCase : Dict = pipe( **A_ , decoder_latents=A_ , super_res_latents=A_ ).images __lowerCAmelCase : Any = self.get_dummy_inputs(A_ , pil_image=A_ ) # Don't pass image, instead pass embedding __lowerCAmelCase : Dict = pipeline_inputs.pop('''image''' ) __lowerCAmelCase : Union[str, Any] = pipe.image_encoder(A_ ).image_embeds __lowerCAmelCase : Optional[int] = pipe( **A_ , decoder_latents=A_ , super_res_latents=A_ , image_embeddings=A_ , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1e-4 @skip_mps def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = torch_device == '''cpu''' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor __lowerCAmelCase : Tuple = 1e-2 self._test_attention_slicing_forward_pass( test_max_difference=A_ , expected_max_diff=A_ ) @skip_mps def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = torch_device == '''cpu''' __lowerCAmelCase : str = True __lowerCAmelCase : int = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] self._test_inference_batch_single_identical( test_max_difference=A_ , relax_max_difference=A_ , additional_params_copy_to_batched_inputs=A_ , ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : List[Any] = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes __lowerCAmelCase : Tuple = [2, 3] self._test_inference_batch_consistent( batch_sizes=A_ , additional_params_copy_to_batched_inputs=A_ , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=A_ ) @skip_mps def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return super().test_save_load_local() @skip_mps def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return super().test_save_load_optional_components() @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png''' ) __lowerCAmelCase : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/unclip/karlo_v1_alpha_cat_variation_fp16.npy''' ) __lowerCAmelCase : Tuple = UnCLIPImageVariationPipeline.from_pretrained( '''kakaobrain/karlo-v1-alpha-image-variations''' , torch_dtype=torch.floataa ) __lowerCAmelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Dict = torch.Generator(device='''cpu''' ).manual_seed(0 ) __lowerCAmelCase : Any = pipeline( A_ , generator=A_ , output_type='''np''' , ) __lowerCAmelCase : List[str] = output.images[0] assert image.shape == (256, 256, 3) assert_mean_pixel_difference(A_ , A_ , 15 )
275
def _lowercase ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __lowerCAmelCase : int = str(bin(lowercase__ ) )[2:] # remove the leading "0b" __lowerCAmelCase : Any = str(bin(lowercase__ ) )[2:] __lowerCAmelCase : List[str] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('''1''' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
275
1
from typing import Union import fire import torch from tqdm import tqdm def _lowercase ( lowercase__ , lowercase__ = "cpu" , lowercase__ = None ): __lowerCAmelCase : str = torch.load(lowercase__ , map_location=lowercase__ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowercase__ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) __lowerCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __lowerCAmelCase : Union[str, Any] = src_path torch.save(lowercase__ , lowercase__ ) if __name__ == "__main__": fire.Fire(convert)
275
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
1
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCAmelCase : Dict = Vector() def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(A_ ) , '''(0,0,0,0,0,1)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(A_ ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Vector([1, 2] ) __lowerCAmelCase : Optional[int] = Vector([1, 2, 3, 4, 5] ) __lowerCAmelCase : Optional[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCAmelCase : str = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : List[Any] = Vector([2, -1, 4] ) # for test of dot product __lowerCAmelCase : Optional[int] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : Any = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , A_ , A_ ) ) , '''(3,4,7)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) __lowerCAmelCase : Optional[Any] = x.copy() self.assertEqual(str(A_ ) , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(A_ ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Dict = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
275
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCamelCase = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
def _lowercase ( lowercase__ , lowercase__ ): if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
275
1
from __future__ import annotations import time import numpy as np _UpperCamelCase = [8, 5, 9, 7] _UpperCamelCase = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCamelCase = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __lowercase : def __init__( self , A_ , A_ , A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : str = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : int = maximum_claim_table def UpperCamelCase__ ( self ) ->list[int]: '''simple docstring''' return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def UpperCamelCase__ ( self ) ->list[int]: '''simple docstring''' return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def UpperCamelCase__ ( self ) ->list[list[int]]: '''simple docstring''' return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(A_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def UpperCamelCase__ ( self ) ->dict[int, list[int]]: '''simple docstring''' return {self.__need().index(A_ ): i for i in self.__need()} def UpperCamelCase__ ( self , **A_ ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : Any = self.__allocated_resources_table __lowerCAmelCase : str = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: __lowerCAmelCase : Tuple = False for each_need in need_list: __lowerCAmelCase : int = True for index, need in enumerate(A_ ): if need > available_resources[index]: __lowerCAmelCase : str = False break if execution: __lowerCAmelCase : Dict = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : int = original_need_index print(f"""Process {process_number + 1} is executing.""" ) # remove the process run from stack need_list.remove(A_ ) # update available/freed resources stack __lowerCAmelCase : str = np.array(A_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(A_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f"""P{self.__allocated_resources_table.index(A_ ) + 1}""" + ''' '''.join(f"""{it:>8}""" for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f"""P{self.__maximum_claim_table.index(A_ ) + 1}""" + ''' '''.join(f"""{it:>8}""" for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(A_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(A_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
275
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
1
import os def _lowercase ( ): __lowerCAmelCase : int = os.path.dirname(os.path.realpath(lowercase__ ) ) __lowerCAmelCase : int = os.path.join(lowercase__ , '''triangle.txt''' ) with open(lowercase__ ) as f: __lowerCAmelCase : Tuple = f.readlines() __lowerCAmelCase : List[Any] = [] for line in triangle: __lowerCAmelCase : Union[str, Any] = [] for number in line.strip().split(''' ''' ): numbers_from_line.append(int(lowercase__ ) ) a.append(lowercase__ ) for i in range(1 , len(lowercase__ ) ): for j in range(len(a[i] ) ): __lowerCAmelCase : Union[str, Any] = a[i - 1][j] if j != len(a[i - 1] ) else 0 __lowerCAmelCase : Optional[int] = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(lowercase__ , lowercase__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class __lowercase : @property def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return self.get_dummy_input() @property def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f"""'{self.block_type}' is not a supported block_type. Set it to 'up', 'mid', or 'down'.""" ) def UpperCamelCase__ ( self , A_=True , A_=False , A_=False , A_=False , ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Dict = 4 __lowerCAmelCase : str = 32 __lowerCAmelCase : Dict = (32, 32) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[str] = torch.device(A_ ) __lowerCAmelCase : Union[str, Any] = (batch_size, num_channels) + sizes __lowerCAmelCase : Any = randn_tensor(A_ , generator=A_ , device=A_ ) __lowerCAmelCase : int = {'''hidden_states''': hidden_states} if include_temb: __lowerCAmelCase : List[Any] = 128 __lowerCAmelCase : List[Any] = randn_tensor((batch_size, temb_channels) , generator=A_ , device=A_ ) if include_res_hidden_states_tuple: __lowerCAmelCase : Any = torch.manual_seed(1 ) __lowerCAmelCase : Union[str, Any] = (randn_tensor(A_ , generator=A_ , device=A_ ),) if include_encoder_hidden_states: __lowerCAmelCase : Optional[int] = floats_tensor((batch_size, 32, 32) ).to(A_ ) if include_skip_sample: __lowerCAmelCase : Any = randn_tensor(((batch_size, 3) + sizes) , generator=A_ , device=A_ ) return dummy_input def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = { '''in_channels''': 32, '''out_channels''': 32, '''temb_channels''': 128, } if self.block_type == "up": __lowerCAmelCase : Optional[Any] = 32 if self.block_type == "mid": init_dict.pop('''out_channels''' ) __lowerCAmelCase : Optional[int] = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Dict = self.prepare_init_args_and_inputs_for_common() __lowerCAmelCase : Optional[int] = self.block_class(**A_ ) unet_block.to(A_ ) unet_block.eval() with torch.no_grad(): __lowerCAmelCase : List[Any] = unet_block(**A_ ) if isinstance(A_ , A_ ): __lowerCAmelCase : Any = output[0] self.assertEqual(output.shape , self.output_shape ) __lowerCAmelCase : Optional[int] = output[0, -1, -3:, -3:] __lowerCAmelCase : int = torch.tensor(A_ ).to(A_ ) assert torch_all_close(output_slice.flatten() , A_ , atol=5e-3 ) @unittest.skipIf(torch_device == '''mps''' , '''Training is not supported in mps''' ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : int = self.prepare_init_args_and_inputs_for_common() __lowerCAmelCase : Any = self.block_class(**A_ ) model.to(A_ ) model.train() __lowerCAmelCase : Tuple = model(**A_ ) if isinstance(A_ , A_ ): __lowerCAmelCase : int = output[0] __lowerCAmelCase : int = torch.device(A_ ) __lowerCAmelCase : Tuple = randn_tensor(output.shape , device=A_ ) __lowerCAmelCase : Optional[Any] = torch.nn.functional.mse_loss(A_ , A_ ) loss.backward()
275
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ConsistencyModelPipeline _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt _UpperCamelCase = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[Any] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def UpperCamelCase__ ( self , A_=False ) ->Dict: '''simple docstring''' if class_cond: __lowerCAmelCase : List[str] = self.dummy_cond_unet else: __lowerCAmelCase : Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->Tuple: '''simple docstring''' if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : str = torch.manual_seed(A_ ) else: __lowerCAmelCase : Dict = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Tuple = self.get_dummy_components() __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : str = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : str = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : int = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Any = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : Union[str, Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : List[str] = 1 __lowerCAmelCase : Dict = None __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , A_=0 , A_=False , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = torch.manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __lowerCAmelCase : List[str] = self.get_fixed_latents(seed=A_ , device=A_ , dtype=A_ , shape=A_ ) __lowerCAmelCase : Union[str, Any] = latents return inputs def UpperCamelCase__ ( self , A_=0 , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->Optional[int]: '''simple docstring''' if type(A_ ) == str: __lowerCAmelCase : int = torch.device(A_ ) __lowerCAmelCase : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Union[str, Any] = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) return latents def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = self.get_inputs() __lowerCAmelCase : Any = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.0_888, 0.0_881, 0.0_666, 0.0_479, 0.0_292, 0.0_195, 0.0_201, 0.0_163, 0.0_254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = self.get_inputs() __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : str = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.0_340, 0.0_152, 0.0_063, 0.0_267, 0.0_221, 0.0_107, 0.0_416, 0.0_186, 0.0_217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_inputs(get_fixed_latents=A_ , device=A_ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.1_875, 0.1_428, 0.1_289, 0.2_151, 0.2_092, 0.1_477, 0.1_877, 0.1_641, 0.1_353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Union[str, Any] = self.get_inputs(get_fixed_latents=A_ , device=A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : int = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : str = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.1_663, 0.1_948, 0.2_275, 0.1_680, 0.1_204, 0.1_245, 0.1_858, 0.1_338, 0.2_095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
275
1
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), } } _UpperCamelCase = { "google/bigbird-roberta-base": 4096, "google/bigbird-roberta-large": 4096, "google/bigbird-base-trivia-itc": 4096, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] _UpperCamelCase = [] def __init__( self , A_ , A_="<unk>" , A_="<s>" , A_="</s>" , A_="<pad>" , A_="[SEP]" , A_="[MASK]" , A_="[CLS]" , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else bos_token __lowerCAmelCase : List[Any] = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else eos_token __lowerCAmelCase : Union[str, Any] = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else unk_token __lowerCAmelCase : Tuple = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else pad_token __lowerCAmelCase : Tuple = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else cls_token __lowerCAmelCase : List[Any] = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase : List[Any] = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else mask_token __lowerCAmelCase : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sep_token=A_ , mask_token=A_ , cls_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Tuple = vocab_file __lowerCAmelCase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.sp_model.get_piece_size() def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[Any] = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.__dict__.copy() __lowerCAmelCase : Dict = None return state def __setstate__( self , A_ ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : str = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase__ ( self , A_ ) ->List[str]: '''simple docstring''' return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.piece_to_id(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.sp_model.IdToPiece(A_ ) return token def UpperCamelCase__ ( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = [] __lowerCAmelCase : Any = '''''' __lowerCAmelCase : Dict = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[int] = True __lowerCAmelCase : List[str] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string.strip() def UpperCamelCase__ ( self , A_ , A_ = False , A_ = None , A_ = True , **A_ , ) ->str: '''simple docstring''' __lowerCAmelCase : List[Any] = kwargs.pop('''use_source_tokenizer''' , A_ ) __lowerCAmelCase : List[str] = self.convert_ids_to_tokens(A_ , skip_special_tokens=A_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 __lowerCAmelCase : Tuple = [] __lowerCAmelCase : List[Any] = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(A_ ) ) __lowerCAmelCase : int = [] sub_texts.append(A_ ) else: current_sub_text.append(A_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(A_ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: __lowerCAmelCase : int = re.sub(R''' (\[(MASK|SEP)\])''' , R'''\1''' , ''' '''.join(A_ ) ) else: __lowerCAmelCase : int = ''''''.join(A_ ) __lowerCAmelCase : List[str] = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: __lowerCAmelCase : List[Any] = self.clean_up_tokenization(A_ ) return clean_text else: return text def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Optional[Any] = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCAmelCase : str = [self.cls_token_id] __lowerCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self , A_ , A_ = None , A_ = False ) ->List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is None: return [1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1] + ([0] * len(A_ )) + [1] def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[int]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = [self.sep_token_id] __lowerCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
275
from collections import deque from .hash_table import HashTable class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->int: '''simple docstring''' super().__init__(*A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(A_ ) __lowerCAmelCase : int = self.values[key] def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return ( sum(self.charge_factor - len(A_ ) for slot in self.values ) / self.size_table * self.charge_factor ) def UpperCamelCase__ ( self , A_ , A_=None ) ->str: '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(A_ ) == 0 ): return key return super()._collision_resolution(A_ , A_ )
275
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = filter(lambda lowercase__ : p.requires_grad , model.parameters() ) __lowerCAmelCase : Any = sum([np.prod(p.size() ) for p in model_parameters] ) return params _UpperCamelCase = logging.getLogger(__name__) def _lowercase ( lowercase__ , lowercase__ ): if metric == "rouge2": __lowerCAmelCase : List[str] = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": __lowerCAmelCase : Union[str, Any] = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": __lowerCAmelCase : str = '''{val_avg_em:.4f}-{step_count}''' else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ''' function.''' ) __lowerCAmelCase : Union[str, Any] = ModelCheckpoint( dirpath=lowercase__ , filename=lowercase__ , monitor=f"""val_{metric}""" , mode='''max''' , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def _lowercase ( lowercase__ , lowercase__ ): return EarlyStopping( monitor=f"""val_{metric}""" , mode='''min''' if '''loss''' in metric else '''max''' , patience=lowercase__ , verbose=lowercase__ , ) class __lowercase (pl.Callback ): def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : Tuple = {f"""lr_group_{i}""": param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(A_ ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ , A_ , A_=True ) ->None: '''simple docstring''' logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase : List[Any] = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']} ) # Log results __lowerCAmelCase : int = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase : int = od / '''test_results.txt''' __lowerCAmelCase : Tuple = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase : List[Any] = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase : int = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=A_ ) generations_file.parent.mkdir(exist_ok=A_ ) with open(A_ , '''a+''' ) as writer: for key in sorted(A_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase : List[Any] = metrics[key] if isinstance(A_ , torch.Tensor ): __lowerCAmelCase : Union[str, Any] = val.item() __lowerCAmelCase : Dict = f"""{key}: {val:.6f}\n""" writer.write(A_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase : List[Any] = '''\n'''.join(metrics['''preds'''] ) generations_file.open('''w+''' ).write(A_ ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' try: __lowerCAmelCase : Optional[int] = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase : List[Any] = pl_module.model.num_parameters() __lowerCAmelCase : Optional[int] = count_trainable_parameters(A_ ) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1e6, '''grad_mp''': n_trainable_pars / 1e6} ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->List[Any]: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(A_ , A_ , '''test''' ) @rank_zero_only def UpperCamelCase__ ( self , A_ , A_ ) ->List[str]: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
275
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
1
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("transformers.models.speecht5") def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): hf_model.apply_weight_norm() __lowerCAmelCase : Any = checkpoint['''input_conv.weight_g'''] __lowerCAmelCase : Dict = checkpoint['''input_conv.weight_v'''] __lowerCAmelCase : List[Any] = checkpoint['''input_conv.bias'''] for i in range(len(config.upsample_rates ) ): __lowerCAmelCase : str = checkpoint[f"""upsamples.{i}.1.weight_g"""] __lowerCAmelCase : str = checkpoint[f"""upsamples.{i}.1.weight_v"""] __lowerCAmelCase : int = checkpoint[f"""upsamples.{i}.1.bias"""] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): __lowerCAmelCase : Optional[int] = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_g"""] __lowerCAmelCase : Tuple = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_v"""] __lowerCAmelCase : int = checkpoint[f"""blocks.{i}.convs1.{j}.1.bias"""] __lowerCAmelCase : Optional[Any] = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_g"""] __lowerCAmelCase : List[str] = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_v"""] __lowerCAmelCase : int = checkpoint[f"""blocks.{i}.convs2.{j}.1.bias"""] __lowerCAmelCase : Union[str, Any] = checkpoint['''output_conv.1.weight_g'''] __lowerCAmelCase : Dict = checkpoint['''output_conv.1.weight_v'''] __lowerCAmelCase : Optional[Any] = checkpoint['''output_conv.1.bias'''] hf_model.remove_weight_norm() @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , ): if config_path is not None: __lowerCAmelCase : int = SpeechTaHifiGanConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Any = SpeechTaHifiGanConfig() __lowerCAmelCase : Optional[Any] = SpeechTaHifiGan(lowercase__ ) __lowerCAmelCase : int = torch.load(lowercase__ ) load_weights(orig_checkpoint['''model''']['''generator'''] , lowercase__ , lowercase__ ) __lowerCAmelCase : Union[str, Any] = np.load(lowercase__ ) __lowerCAmelCase : int = stats[0].reshape(-1 ) __lowerCAmelCase : Any = stats[1].reshape(-1 ) __lowerCAmelCase : Optional[Any] = torch.from_numpy(lowercase__ ).float() __lowerCAmelCase : str = torch.from_numpy(lowercase__ ).float() model.save_pretrained(lowercase__ ) if repo_id: print('''Pushing to the hub...''' ) model.push_to_hub(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
275
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=False , A_=True , A_="None" , A_=3 , A_=4 , A_=None , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = parent __lowerCAmelCase : List[str] = batch_size __lowerCAmelCase : Dict = seq_length __lowerCAmelCase : List[Any] = is_training __lowerCAmelCase : List[Any] = use_input_mask __lowerCAmelCase : Optional[int] = use_token_type_ids __lowerCAmelCase : Tuple = use_labels __lowerCAmelCase : str = vocab_size __lowerCAmelCase : int = hidden_size __lowerCAmelCase : Any = num_hidden_layers __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : int = hidden_act __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Union[str, Any] = type_sequence_label_size __lowerCAmelCase : Optional[int] = initializer_range __lowerCAmelCase : int = num_labels __lowerCAmelCase : int = num_choices __lowerCAmelCase : List[str] = relative_attention __lowerCAmelCase : Union[str, Any] = position_biased_input __lowerCAmelCase : int = pos_att_type __lowerCAmelCase : List[Any] = scope def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : int = None if self.use_input_mask: __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __lowerCAmelCase : List[str] = None if self.use_token_type_ids: __lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : int = None __lowerCAmelCase : List[str] = None if self.use_labels: __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_config() __lowerCAmelCase : Dict = 300 return config def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = DebertaModel(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ )[0] __lowerCAmelCase : Any = model(A_ , token_type_ids=A_ )[0] __lowerCAmelCase : List[str] = model(A_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = DebertaForMaskedLM(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = self.num_labels __lowerCAmelCase : Tuple = DebertaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Optional[int] = DebertaForTokenClassification(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Tuple = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = DebertaForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : int = model( A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ) : Tuple = config_and_inputs __lowerCAmelCase : Tuple = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowercase (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : int = DebertaModelTester(self ) __lowerCAmelCase : List[Any] = ConfigTester(self , config_class=A_ , hidden_size=37 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*A_ ) @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[int] = DebertaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowercase (unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' pass @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = DebertaModel.from_pretrained('''microsoft/deberta-base''' ) __lowerCAmelCase : Tuple = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase : Optional[int] = model(A_ , attention_mask=A_ )[0] # compare the actual values for a slice. __lowerCAmelCase : Optional[Any] = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , A_ , atol=1e-4 ) , f"""{output[:, 1:4, 1:4]}""" )
275
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowercase (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = StableDiffusionInstructPixaPixPipeline _UpperCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} _UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS _UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) __lowerCAmelCase : List[Any] = PNDMScheduler(skip_prk_steps=A_ ) torch.manual_seed(0 ) __lowerCAmelCase : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __lowerCAmelCase : Optional[Any] = CLIPTextModel(A_ ) __lowerCAmelCase : Optional[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) __lowerCAmelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_ ) ).to(A_ ) __lowerCAmelCase : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase : Any = Image.fromarray(np.uinta(A_ ) ).convert('''RGB''' ) if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : Dict = torch.manual_seed(A_ ) else: __lowerCAmelCase : int = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : List[Any] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''image_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Dict = self.get_dummy_components() __lowerCAmelCase : List[str] = StableDiffusionInstructPixaPixPipeline(**A_ ) __lowerCAmelCase : int = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[str] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Union[str, Any] = sd_pipe(**A_ ).images __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.7_526, 0.3_750, 0.4_547, 0.6_117, 0.5_866, 0.5_016, 0.4_327, 0.5_642, 0.4_815] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : str = self.get_dummy_components() __lowerCAmelCase : List[str] = StableDiffusionInstructPixaPixPipeline(**A_ ) __lowerCAmelCase : List[str] = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : int = self.get_dummy_inputs(A_ ) __lowerCAmelCase : int = '''french fries''' __lowerCAmelCase : str = sd_pipe(**A_ , negative_prompt=A_ ) __lowerCAmelCase : List[Any] = output.images __lowerCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([0.7_511, 0.3_642, 0.4_553, 0.6_236, 0.5_797, 0.5_013, 0.4_343, 0.5_611, 0.4_831] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Any = self.get_dummy_components() __lowerCAmelCase : int = StableDiffusionInstructPixaPixPipeline(**A_ ) __lowerCAmelCase : Optional[int] = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Tuple = [inputs['''prompt''']] * 2 __lowerCAmelCase : List[Any] = np.array(inputs['''image'''] ).astype(np.floataa ) / 255.0 __lowerCAmelCase : Tuple = torch.from_numpy(A_ ).unsqueeze(0 ).to(A_ ) __lowerCAmelCase : int = image / 2 + 0.5 __lowerCAmelCase : int = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase : Optional[int] = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase : int = sd_pipe(**A_ ).images __lowerCAmelCase : List[Any] = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __lowerCAmelCase : Any = np.array([0.5_812, 0.5_748, 0.5_222, 0.5_908, 0.5_695, 0.7_174, 0.6_804, 0.5_523, 0.5_579] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : Any = EulerAncestralDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' ) __lowerCAmelCase : int = StableDiffusionInstructPixaPixPipeline(**A_ ) __lowerCAmelCase : Dict = sd_pipe.to(A_ ) sd_pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[int] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Optional[Any] = sd_pipe(**A_ ).images __lowerCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = [round(A_ , 4 ) for x in image_slice.flatten().tolist()] print(''','''.join([str(A_ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[str] = np.array([0.7_417, 0.3_842, 0.4_732, 0.5_776, 0.5_891, 0.5_139, 0.4_052, 0.5_673, 0.4_986] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : Any = StableDiffusionInstructPixaPixPipeline(**A_ ) __lowerCAmelCase : Dict = VaeImageProcessor(do_resize=A_ , do_normalize=A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[str] = pipe(**self.get_dummy_inputs_by_type(A_ , input_image_type='''pt''' ) )[0] __lowerCAmelCase : Any = components['''vae'''] __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs_by_type(A_ , input_image_type='''pt''' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase : Any = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase : Optional[int] = pipe(**A_ )[0] __lowerCAmelCase : Optional[Any] = np.abs(out - out_latents_inputs ).max() self.assertLess(A_ , 1e-4 , '''passing latents as image input generate different result from passing image''' ) @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , A_=0 ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.manual_seed(A_ ) __lowerCAmelCase : Optional[int] = load_image( '''https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg''' ) __lowerCAmelCase : Optional[int] = { '''prompt''': '''turn him into a cyborg''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''image_guidance_scale''': 1.0, '''output_type''': '''numpy''', } return inputs def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=A_ ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() __lowerCAmelCase : int = self.get_inputs() __lowerCAmelCase : Union[str, Any] = pipe(**A_ ).images __lowerCAmelCase : str = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCAmelCase : Optional[Any] = np.array([0.5_902, 0.6_015, 0.6_027, 0.5_983, 0.6_092, 0.6_061, 0.5_765, 0.5_785, 0.5_555] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=A_ ) __lowerCAmelCase : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() __lowerCAmelCase : Optional[Any] = self.get_inputs() __lowerCAmelCase : Dict = pipe(**A_ ).images __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCAmelCase : Optional[Any] = np.array([0.6_578, 0.6_817, 0.6_972, 0.6_761, 0.6_856, 0.6_916, 0.6_428, 0.6_516, 0.6_301] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=A_ ) __lowerCAmelCase : List[Any] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() __lowerCAmelCase : Optional[Any] = self.get_inputs() __lowerCAmelCase : Union[str, Any] = pipe(**A_ ).images __lowerCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCAmelCase : Optional[Any] = np.array([0.3_828, 0.3_834, 0.3_818, 0.3_792, 0.3_865, 0.3_752, 0.3_792, 0.3_847, 0.3_753] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = 0 def callback_fn(A_ , A_ , A_ ) -> None: __lowerCAmelCase : Union[str, Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase : Any = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCAmelCase : List[str] = latents[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([-0.2_463, -0.4_644, -0.9_756, 1.5_176, 1.4_414, 0.7_866, 0.9_897, 0.8_521, 0.7_983] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase : int = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCAmelCase : List[str] = latents[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([-0.2_644, -0.4_626, -0.9_653, 1.5_176, 1.4_551, 0.7_686, 0.9_805, 0.8_452, 0.8_115] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase : List[str] = False __lowerCAmelCase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=A_ , torch_dtype=torch.floataa ) __lowerCAmelCase : List[Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() __lowerCAmelCase : Optional[int] = self.get_inputs() pipe(**A_ , callback=A_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=A_ , torch_dtype=torch.floataa ) __lowerCAmelCase : str = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase : Union[str, Any] = self.get_inputs() __lowerCAmelCase : int = pipe(**A_ ) __lowerCAmelCase : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase : List[str] = inputs['''image'''].resize((504, 504) ) __lowerCAmelCase : Tuple = '''timbrooks/instruct-pix2pix''' __lowerCAmelCase : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( A_ , safety_checker=A_ , ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) pipe.enable_attention_slicing() __lowerCAmelCase : List[Any] = pipe(**A_ ) __lowerCAmelCase : Optional[int] = output.images[0] __lowerCAmelCase : Any = image[255:258, 383:386, -1] assert image.shape == (504, 504, 3) __lowerCAmelCase : int = np.array([0.2_726, 0.2_529, 0.2_664, 0.2_655, 0.2_641, 0.2_642, 0.2_591, 0.2_649, 0.2_590] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
275
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : str = [] for rt in rc.restypes: __lowerCAmelCase : List[Any] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase : List[str] = {name: i for i, name in enumerate(lowercase__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) __lowerCAmelCase : List[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Optional[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Tuple = torch.tensor( lowercase__ , dtype=torch.floataa , device=protein['''aatype'''].device , ) __lowerCAmelCase : List[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : int = residx_atomaa_mask __lowerCAmelCase : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase : int = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase : str = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase : Optional[int] = rc.restype_atoa[restype_letter] __lowerCAmelCase : Optional[Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase : str = rc.atom_order[atom_name] __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : Any = residx_atomaa_mask return protein def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = tree_map(lambda lowercase__ : torch.tensor(lowercase__ , device=batch['''aatype'''].device ) , lowercase__ , np.ndarray ) __lowerCAmelCase : Tuple = tensor_tree_map(lambda lowercase__ : np.array(lowercase__ ) , make_atomaa_masks(lowercase__ ) ) return out
275
1
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , ) ->int: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 18, '''width''': 18} __lowerCAmelCase : Dict = parent __lowerCAmelCase : int = batch_size __lowerCAmelCase : Union[str, Any] = num_channels __lowerCAmelCase : Tuple = image_size __lowerCAmelCase : List[str] = min_resolution __lowerCAmelCase : Dict = max_resolution __lowerCAmelCase : Any = do_resize __lowerCAmelCase : Optional[Any] = size __lowerCAmelCase : str = do_normalize __lowerCAmelCase : Optional[int] = image_mean __lowerCAmelCase : Optional[int] = image_std def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = DPTImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[int] = DPTImageProcessingTester(self ) @property def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , '''image_mean''' ) ) self.assertTrue(hasattr(A_ , '''image_std''' ) ) self.assertTrue(hasattr(A_ , '''do_normalize''' ) ) self.assertTrue(hasattr(A_ , '''do_resize''' ) ) self.assertTrue(hasattr(A_ , '''size''' ) ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) __lowerCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input __lowerCAmelCase : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __lowerCAmelCase : Any = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input __lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __lowerCAmelCase : Optional[Any] = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input __lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __lowerCAmelCase : Optional[Any] = image_processing(A_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , )
275
def _lowercase ( lowercase__ ): if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) __lowerCAmelCase : int = sorted(string.lower() ) return len(lowercase__ ) == len(set(lowercase__ ) ) if __name__ == "__main__": _UpperCamelCase = input("Enter a string ").strip() _UpperCamelCase = is_isogram(input_str) print(F"{input_str} is {'an' if isogram else 'not an'} isogram.")
275
1
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = 1 / 255 , A_ = True , A_ = 8 , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Tuple = do_rescale __lowerCAmelCase : int = rescale_factor __lowerCAmelCase : List[Any] = do_pad __lowerCAmelCase : Any = pad_size def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ ) ->np.ndarray: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->int: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : str = get_image_size(A_ ) __lowerCAmelCase : Tuple = (old_height // size + 1) * size - old_height __lowerCAmelCase : Tuple = (old_width // size + 1) * size - old_width return pad(A_ , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Dict = do_pad if do_pad is not None else self.do_pad __lowerCAmelCase : str = pad_size if pad_size is not None else self.pad_size __lowerCAmelCase : Any = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Union[str, Any] = [to_numpy_array(A_ ) for image in images] if do_rescale: __lowerCAmelCase : Optional[int] = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_pad: __lowerCAmelCase : List[str] = [self.pad(A_ , size=A_ ) for image in images] __lowerCAmelCase : Optional[int] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Any = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ )
275
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
1
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def _lowercase ( lowercase__ , lowercase__=False ): __lowerCAmelCase : Union[str, Any] = OmegaConf.load(lowercase__ ) if display: print(yaml.dump(OmegaConf.to_container(lowercase__ ) ) ) return config def _lowercase ( lowercase__ , lowercase__=None , lowercase__=None ): if conf_path is None: __lowerCAmelCase : Tuple = '''./model_checkpoints/vqgan_only.yaml''' __lowerCAmelCase : Tuple = load_config(lowercase__ , display=lowercase__ ) __lowerCAmelCase : Optional[Any] = VQModel(**config.model.params ) if ckpt_path is None: __lowerCAmelCase : List[str] = '''./model_checkpoints/vqgan_only.pt''' __lowerCAmelCase : List[str] = torch.load(lowercase__ , map_location=lowercase__ ) if ".ckpt" in ckpt_path: __lowerCAmelCase : Tuple = sd['''state_dict'''] model.load_state_dict(lowercase__ , strict=lowercase__ ) model.to(lowercase__ ) del sd return model def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : str = model.encode(lowercase__ ) print(f"""VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}""" ) __lowerCAmelCase : Optional[int] = model.decode(lowercase__ ) return xrec def _lowercase ( lowercase__ , lowercase__=False ): __lowerCAmelCase, __lowerCAmelCase : str = string.rsplit('''.''' , 1 ) if reload: __lowerCAmelCase : Any = importlib.import_module(lowercase__ ) importlib.reload(lowercase__ ) return getattr(importlib.import_module(lowercase__ , package=lowercase__ ) , cls ) def _lowercase ( lowercase__ ): if "target" not in config: raise KeyError('''Expected key `target` to instantiate.''' ) return get_obj_from_str(config['''target'''] )(**config.get('''params''' , {} ) ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=True , lowercase__=True ): __lowerCAmelCase : Tuple = instantiate_from_config(lowercase__ ) if sd is not None: model.load_state_dict(lowercase__ ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): # load the specified checkpoint if ckpt: __lowerCAmelCase : List[Any] = torch.load(lowercase__ , map_location='''cpu''' ) __lowerCAmelCase : Dict = pl_sd['''global_step'''] print(f"""loaded model from global step {global_step}.""" ) else: __lowerCAmelCase : Any = {'''state_dict''': None} __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = load_model_from_config(config.model , pl_sd['''state_dict'''] , gpu=lowercase__ , eval_mode=lowercase__ )['''model'''] return model, global_step
275
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = np.full((len(lowercase__ ), sequence_length, 2) , lowercase__ ) else: __lowerCAmelCase : Optional[int] = np.full((len(lowercase__ ), sequence_length) , lowercase__ ) for i, tensor in enumerate(lowercase__ ): if padding_side == "right": if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : int = tensor[:sequence_length] else: if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = ord(lowercase__ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True __lowerCAmelCase : int = unicodedata.category(lowercase__ ) if cat.startswith('''P''' ): return True return False @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = True _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = -100 _UpperCamelCase = "pt" def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCAmelCase : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCAmelCase : List[Any] = self.tokenizer.pad( A_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCAmelCase : Dict = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCAmelCase : Optional[int] = self.tokenizer.padding_side if padding_side == "right": __lowerCAmelCase : Any = [ list(A_ ) + [self.label_pad_token_id] * (sequence_length - len(A_ )) for label in labels ] else: __lowerCAmelCase : Optional[int] = [ [self.label_pad_token_id] * (sequence_length - len(A_ )) + list(A_ ) for label in labels ] __lowerCAmelCase : Tuple = [feature['''ner_tags'''] for feature in features] __lowerCAmelCase : List[Any] = padding_tensor(A_ , -1 , A_ , A_ ) __lowerCAmelCase : Optional[int] = [feature['''original_entity_spans'''] for feature in features] __lowerCAmelCase : Any = padding_tensor(A_ , (-1, -1) , A_ , A_ ) __lowerCAmelCase : Optional[Any] = {k: torch.tensor(A_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
275
1
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->None: '''simple docstring''' warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
275
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase (unittest.TestCase ): _UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Tuple = (3, 32, 128) __lowerCAmelCase : List[str] = tempfile.mkdtemp() # fmt: off __lowerCAmelCase : List[str] = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on __lowerCAmelCase : Optional[int] = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) __lowerCAmelCase : Union[str, Any] = { '''do_normalize''': False, '''do_resize''': True, '''image_processor_type''': '''ViTImageProcessor''', '''resample''': 3, '''size''': {'''height''': 32, '''width''': 128}, } __lowerCAmelCase : Optional[Any] = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(A_ , A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) __lowerCAmelCase : str = Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) return image_input def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Union[str, Any] = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : int = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) __lowerCAmelCase : int = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.get_image_processor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Optional[int] = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = image_processor(A_ , return_tensors='''np''' ) __lowerCAmelCase : Tuple = processor(images=A_ , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Union[str, Any] = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Any = '''test''' __lowerCAmelCase : Dict = processor(text=A_ ) __lowerCAmelCase : str = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : str = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = '''test''' __lowerCAmelCase : int = self.prepare_image_inputs() __lowerCAmelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''pixel_values''', '''labels'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Optional[int] = processor.char_decode(A_ ) __lowerCAmelCase : Tuple = tokenizer.batch_decode(A_ ) __lowerCAmelCase : Any = [seq.replace(''' ''' , '''''' ) for seq in decoded_tok] self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = self.prepare_image_inputs() __lowerCAmelCase : List[Any] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_image_processor() __lowerCAmelCase : List[str] = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = torch.randn(1 , 27 , 38 ) __lowerCAmelCase : Optional[int] = torch.randn(1 , 27 , 5_0257 ) __lowerCAmelCase : Optional[Any] = torch.randn(1 , 27 , 3_0522 ) __lowerCAmelCase : List[str] = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['''generated_text''', '''scores''', '''char_preds''', '''bpe_preds''', '''wp_preds'''] )
275
1
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py _UpperCamelCase = "." if __name__ == "__main__": _UpperCamelCase = os.path.join(REPO_PATH, "utils/documentation_tests.txt") _UpperCamelCase = [] _UpperCamelCase = [] with open(doctest_file_path) as fp: for line in fp: _UpperCamelCase = line.strip() _UpperCamelCase = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: _UpperCamelCase = "\n".join(non_existent_paths) raise ValueError(F"`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}") if all_paths != sorted(all_paths): raise ValueError("Files in `utils/documentation_tests.txt` are not in alphabetical order.")
275
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = self.dummy_uncond_unet __lowerCAmelCase : Any = PNDMScheduler() __lowerCAmelCase : Dict = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' ).images __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''google/ddpm-cifar10-32''' __lowerCAmelCase : Union[str, Any] = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : int = PNDMScheduler() __lowerCAmelCase : Any = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
1
import numpy as np def _lowercase ( lowercase__ ): return 1 / (1 + np.exp(-vector )) if __name__ == "__main__": import doctest doctest.testmod()
275
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
1
import math import os import sys def _lowercase ( lowercase__ ): __lowerCAmelCase : Any = '''''' try: with open(lowercase__ , '''rb''' ) as binary_file: __lowerCAmelCase : int = binary_file.read() for dat in data: __lowerCAmelCase : Union[str, Any] = f"""{dat:08b}""" result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): lexicon.pop(lowercase__ ) __lowerCAmelCase : str = last_match_id if math.loga(lowercase__ ).is_integer(): for curr_key in lexicon: __lowerCAmelCase : Optional[int] = '''0''' + lexicon[curr_key] __lowerCAmelCase : Any = bin(lowercase__ )[2:] def _lowercase ( lowercase__ ): __lowerCAmelCase : int = {'''0''': '''0''', '''1''': '''1'''} __lowerCAmelCase, __lowerCAmelCase : str = '''''', '''''' __lowerCAmelCase : Tuple = len(lowercase__ ) for i in range(len(lowercase__ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue __lowerCAmelCase : Optional[int] = lexicon[curr_string] result += last_match_id add_key_to_lexicon(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) index += 1 __lowerCAmelCase : Optional[int] = '''''' while curr_string != "" and curr_string not in lexicon: curr_string += "0" if curr_string != "": __lowerCAmelCase : Optional[int] = lexicon[curr_string] result += last_match_id return result def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = os.path.getsize(lowercase__ ) __lowerCAmelCase : Optional[Any] = bin(lowercase__ )[2:] __lowerCAmelCase : Any = len(lowercase__ ) return "0" * (length_length - 1) + file_length_binary + compressed def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Tuple = 8 try: with open(lowercase__ , '''wb''' ) as opened_file: __lowerCAmelCase : Dict = [ to_write[i : i + byte_length] for i in range(0 , len(lowercase__ ) , lowercase__ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array: opened_file.write(int(lowercase__ , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = read_file_binary(lowercase__ ) __lowerCAmelCase : List[Any] = compress_data(lowercase__ ) __lowerCAmelCase : int = add_file_length(lowercase__ , lowercase__ ) write_file_binary(lowercase__ , lowercase__ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
275
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _UpperCamelCase = { "AI-Sweden/gpt-sw3-126m": 2048, "AI-Sweden/gpt-sw3-350m": 2048, "AI-Sweden/gpt-sw3-1.6b": 2048, "AI-Sweden/gpt-sw3-6.7b": 2048, "AI-Sweden/gpt-sw3-20b": 2048, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_ , A_=False , A_=False , A_=False , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCAmelCase : int = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCAmelCase : Union[str, Any] = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCAmelCase : str = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCAmelCase : Any = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token __lowerCAmelCase : int = eos_token if bos_token is None else bos_token else: __lowerCAmelCase : Optional[int] = '''<pad>''' if pad_token is None else pad_token __lowerCAmelCase : List[str] = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=A_ , remove_space=A_ , keep_accents=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Union[str, Any] = do_lower_case __lowerCAmelCase : Union[str, Any] = remove_space __lowerCAmelCase : int = keep_accents __lowerCAmelCase : Union[str, Any] = vocab_file __lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # Used for whitespace normalization in input texts # fmt : off __lowerCAmelCase : List[Any] = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCAmelCase : int = re.compile( f"""[{"".join(map(A_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.__dict__.copy() __lowerCAmelCase : List[Any] = None return state def __setstate__( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.non_printing_characters_re.sub('''''' , A_ ) # Normalize whitespaces __lowerCAmelCase : List[str] = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCAmelCase : Tuple = unicodedata.normalize('''NFC''' , A_ ) return text def UpperCamelCase__ ( self , A_ , **A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = self.preprocess_text(A_ ) return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' return self.sp_model.PieceToId(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.IdToPiece(A_ ) @staticmethod def UpperCamelCase__ ( A_ ) ->str: '''simple docstring''' return out_string def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = [] __lowerCAmelCase : Tuple = '''''' __lowerCAmelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string def UpperCamelCase__ ( self ) ->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : str = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Any = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = False ) ->Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[Any] = self.preprocess_text(A_ ) __lowerCAmelCase : Dict = self.sp_model.encode(A_ ) else: __lowerCAmelCase : Dict = [self.preprocess_text(A_ ) for t in text] __lowerCAmelCase : Optional[int] = self.sp_model.encode(A_ ) if return_tensors is True or return_tensors == "pt": __lowerCAmelCase : Tuple = torch.tensor(A_ ) return token_ids def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.decode(A_ ) def UpperCamelCase__ ( self , A_ ) ->List[int]: '''simple docstring''' __lowerCAmelCase : int = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCAmelCase : Any = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(A_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=A_ )
275
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_ = None , A_ = None , A_ = True , A_ = None , A_ = False , A_ = None , A_ = True , A_ = "arrow" , **A_ , ) ->str: '''simple docstring''' super().__init__( split=A_ , features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , **A_ , ) __lowerCAmelCase : Optional[int] = load_from_cache_file __lowerCAmelCase : Dict = file_format __lowerCAmelCase : int = Spark( df=A_ , features=A_ , cache_dir=A_ , working_dir=A_ , **A_ , ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) __lowerCAmelCase : Optional[Any] = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=A_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
275
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """table-transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) ->Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowerCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(A_ , A_ ): __lowerCAmelCase : int = backbone_config.get('''model_type''' ) __lowerCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __lowerCAmelCase : Any = config_class.from_dict(A_ ) # set timm attributes to None __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = None, None, None __lowerCAmelCase : Tuple = use_timm_backbone __lowerCAmelCase : Optional[Any] = backbone_config __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = num_queries __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_ffn_dim __lowerCAmelCase : Optional[int] = encoder_layers __lowerCAmelCase : List[str] = encoder_attention_heads __lowerCAmelCase : str = decoder_ffn_dim __lowerCAmelCase : Union[str, Any] = decoder_layers __lowerCAmelCase : Any = decoder_attention_heads __lowerCAmelCase : Optional[int] = dropout __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : Tuple = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : List[str] = init_std __lowerCAmelCase : Tuple = init_xavier_std __lowerCAmelCase : Any = encoder_layerdrop __lowerCAmelCase : List[Any] = decoder_layerdrop __lowerCAmelCase : Optional[Any] = encoder_layers __lowerCAmelCase : Optional[Any] = auxiliary_loss __lowerCAmelCase : Optional[Any] = position_embedding_type __lowerCAmelCase : Tuple = backbone __lowerCAmelCase : Any = use_pretrained_backbone __lowerCAmelCase : int = dilation # Hungarian matcher __lowerCAmelCase : Dict = class_cost __lowerCAmelCase : List[str] = bbox_cost __lowerCAmelCase : int = giou_cost # Loss coefficients __lowerCAmelCase : Optional[Any] = mask_loss_coefficient __lowerCAmelCase : Tuple = dice_loss_coefficient __lowerCAmelCase : int = bbox_loss_coefficient __lowerCAmelCase : List[Any] = giou_loss_coefficient __lowerCAmelCase : int = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.d_model class __lowercase (_UpperCAmelCase ): _UpperCamelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def UpperCamelCase__ ( self ) ->float: '''simple docstring''' return 1e-5 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 12
275
1
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = RoFormerTokenizer _UpperCamelCase = RoFormerTokenizerFast _UpperCamelCase = True _UpperCamelCase = True def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' super().setUp() def UpperCamelCase__ ( self , **A_ ) ->str: '''simple docstring''' return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Union[str, Any]: '''simple docstring''' return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = '''永和服装饰品有限公司,今天天气非常好''' __lowerCAmelCase : int = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase, __lowerCAmelCase : str = self.get_chinese_input_output_texts() __lowerCAmelCase : Optional[int] = tokenizer.tokenize(A_ ) self.assertListEqual(A_ , output_text.split() ) __lowerCAmelCase : Union[str, Any] = tokens + [tokenizer.unk_token] __lowerCAmelCase : List[Any] = [2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = self.get_rust_tokenizer() __lowerCAmelCase, __lowerCAmelCase : Dict = self.get_chinese_input_output_texts() __lowerCAmelCase : Optional[int] = tokenizer.tokenize(A_ ) self.assertListEqual(A_ , output_text.split() ) __lowerCAmelCase : str = tokens + [tokenizer.unk_token] __lowerCAmelCase : Optional[Any] = [2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' pass def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' pass def UpperCamelCase__ ( self ) ->int: '''simple docstring''' pass
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
import numpy as np _UpperCamelCase = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class __lowercase : def __init__( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = np.array(A_ ) def UpperCamelCase__ ( self , A_ ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Optional[int] = np.where(letter == self.SQUARE ) __lowerCAmelCase : str = np.concatenate([indexa + 1, indexa + 1] ) return indexes def UpperCamelCase__ ( self , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.SQUARE[indexa - 1, indexa - 1] return letter def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = message.lower() __lowerCAmelCase : Any = message.replace(''' ''' , '''''' ) __lowerCAmelCase : Optional[Any] = message.replace('''j''' , '''i''' ) __lowerCAmelCase : Optional[Any] = np.empty((2, len(A_ )) ) for letter_index in range(len(A_ ) ): __lowerCAmelCase : List[str] = self.letter_to_numbers(message[letter_index] ) __lowerCAmelCase : List[Any] = numbers[0] __lowerCAmelCase : List[Any] = numbers[1] __lowerCAmelCase : Tuple = first_step.reshape(2 * len(A_ ) ) __lowerCAmelCase : Tuple = '''''' for numbers_index in range(len(A_ ) ): __lowerCAmelCase : Any = int(second_step[numbers_index * 2] ) __lowerCAmelCase : Tuple = int(second_step[(numbers_index * 2) + 1] ) __lowerCAmelCase : Any = self.numbers_to_letter(A_ , A_ ) __lowerCAmelCase : Any = encoded_message + letter return encoded_message def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[Any] = message.lower() message.replace(''' ''' , '''''' ) __lowerCAmelCase : Tuple = np.empty(2 * len(A_ ) ) for letter_index in range(len(A_ ) ): __lowerCAmelCase : Union[str, Any] = self.letter_to_numbers(message[letter_index] ) __lowerCAmelCase : List[Any] = numbers[0] __lowerCAmelCase : str = numbers[1] __lowerCAmelCase : int = first_step.reshape((2, len(A_ )) ) __lowerCAmelCase : Optional[Any] = '''''' for numbers_index in range(len(A_ ) ): __lowerCAmelCase : int = int(second_step[0, numbers_index] ) __lowerCAmelCase : Tuple = int(second_step[1, numbers_index] ) __lowerCAmelCase : Any = self.numbers_to_letter(A_ , A_ ) __lowerCAmelCase : Tuple = decoded_message + letter return decoded_message
275
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = {} with open(lowercase__ , '''r''' ) as file: for line_number, line in enumerate(lowercase__ ): __lowerCAmelCase : Any = line.strip() if line: __lowerCAmelCase : Dict = line.split() __lowerCAmelCase : str = line_number __lowerCAmelCase : List[str] = words[0] __lowerCAmelCase : Any = value return result def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[Any] = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : Tuple = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : List[Any] = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : str = getattr(lowercase__ , lowercase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCAmelCase : Dict = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : str = shape_pointer.shape # let's reduce dimension __lowerCAmelCase : Any = value[0] else: __lowerCAmelCase : str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __lowerCAmelCase : List[str] = value elif weight_type == "weight_v": __lowerCAmelCase : int = value elif weight_type == "bias": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : int = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : Tuple = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCAmelCase : List[str] = '''.'''.join([key, hf_param_name] ) else: __lowerCAmelCase : Optional[int] = key __lowerCAmelCase : Union[str, Any] = value if '''lm_head''' in full_key else value[0] _UpperCamelCase = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None ): __lowerCAmelCase : Any = False for key, mapped_key in MAPPING.items(): __lowerCAmelCase : Tuple = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : List[str] = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : List[Any] = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : Any = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCAmelCase : int = '''weight''' else: __lowerCAmelCase : Any = None if hf_dict is not None: rename_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return is_used return is_used def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[Any] = fairseq_model.state_dict() __lowerCAmelCase : Tuple = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : int = True else: __lowerCAmelCase : Dict = load_wavaveca_layer(lowercase__ , lowercase__ , lowercase__ ) if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : List[str] = name.split('''.''' ) __lowerCAmelCase : Any = int(items[0] ) __lowerCAmelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=True , lowercase__=False ): if config_path is not None: __lowerCAmelCase : Union[str, Any] = WavaVecaConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Optional[int] = WavaVecaConfig() if is_seq_class: __lowerCAmelCase : Optional[Any] = read_txt_into_dict(lowercase__ ) __lowerCAmelCase : int = idalabel __lowerCAmelCase : Optional[int] = WavaVecaForSequenceClassification(lowercase__ ) __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) feature_extractor.save_pretrained(lowercase__ ) elif is_finetuned: if dict_path: __lowerCAmelCase : List[str] = Dictionary.load(lowercase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCAmelCase : List[Any] = target_dict.pad_index __lowerCAmelCase : List[Any] = target_dict.bos_index __lowerCAmelCase : Optional[int] = target_dict.eos_index __lowerCAmelCase : Any = len(target_dict.symbols ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''vocab.json''' ) if not os.path.isdir(lowercase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(lowercase__ ) ) return os.makedirs(lowercase__ , exist_ok=lowercase__ ) __lowerCAmelCase : Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : int = 1 with open(lowercase__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = WavaVecaCTCTokenizer( lowercase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=lowercase__ , ) __lowerCAmelCase : List[str] = True if config.feat_extract_norm == '''layer''' else False __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=lowercase__ , tokenizer=lowercase__ ) processor.save_pretrained(lowercase__ ) __lowerCAmelCase : str = WavaVecaForCTC(lowercase__ ) else: __lowerCAmelCase : Any = WavaVecaForPreTraining(lowercase__ ) if is_finetuned or is_seq_class: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCAmelCase : Union[str, Any] = argparse.Namespace(task='''audio_pretraining''' ) __lowerCAmelCase : str = fairseq.tasks.setup_task(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowercase__ ) __lowerCAmelCase : int = model[0].eval() recursively_load_weights(lowercase__ , lowercase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
275
1
import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=13 , A_=30 , A_=2 , A_=3 , A_=True , A_=True , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=10 , A_=0.02 , ) ->int: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Dict = image_size __lowerCAmelCase : Optional[Any] = patch_size __lowerCAmelCase : Dict = num_channels __lowerCAmelCase : Union[str, Any] = is_training __lowerCAmelCase : Optional[int] = use_labels __lowerCAmelCase : int = hidden_size __lowerCAmelCase : Tuple = num_hidden_layers __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Union[str, Any] = intermediate_size __lowerCAmelCase : Optional[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : str = attention_probs_dropout_prob __lowerCAmelCase : str = type_sequence_label_size __lowerCAmelCase : List[str] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : Union[str, Any] = (image_size // patch_size) ** 2 __lowerCAmelCase : Optional[int] = num_patches + 1 def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase : Any = ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) return config, pixel_values def UpperCamelCase__ ( self , A_ , A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = FlaxViTModel(config=A_ ) __lowerCAmelCase : Tuple = model(A_ ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : Tuple = (self.image_size, self.image_size) __lowerCAmelCase : Optional[Any] = (self.patch_size, self.patch_size) __lowerCAmelCase : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def UpperCamelCase__ ( self , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.type_sequence_label_size __lowerCAmelCase : Optional[Any] = FlaxViTForImageClassification(config=A_ ) __lowerCAmelCase : Any = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCAmelCase : Dict = 1 __lowerCAmelCase : Union[str, Any] = FlaxViTForImageClassification(A_ ) __lowerCAmelCase : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase : Dict = model(A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : int = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ), ( __lowerCAmelCase ), ) : Optional[Any] = config_and_inputs __lowerCAmelCase : Optional[int] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = FlaxViTModelTester(self ) __lowerCAmelCase : Optional[Any] = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : Union[str, Any] = model_class(A_ ) __lowerCAmelCase : Tuple = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase : Optional[int] = [*signature.parameters.keys()] __lowerCAmelCase : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __lowerCAmelCase : List[Any] = self._prepare_for_class(A_ , A_ ) __lowerCAmelCase : Optional[Any] = model_class(A_ ) @jax.jit def model_jitted(A_ , **A_ ): return model(pixel_values=A_ , **A_ ) with self.subTest('''JIT Enabled''' ): __lowerCAmelCase : List[Any] = model_jitted(**A_ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): __lowerCAmelCase : Any = model_jitted(**A_ ).to_tuple() self.assertEqual(len(A_ ) , len(A_ ) ) for jitted_output, output in zip(A_ , A_ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_class_name in self.all_model_classes: __lowerCAmelCase : Union[str, Any] = model_class_name.from_pretrained('''google/vit-base-patch16-224''' ) __lowerCAmelCase : List[str] = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(A_ )
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
from __future__ import annotations def _lowercase ( lowercase__ ): # This function is recursive __lowerCAmelCase : str = len(lowercase__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else __lowerCAmelCase : List[Any] = array[0] __lowerCAmelCase : List[Any] = False __lowerCAmelCase : int = 1 __lowerCAmelCase : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: __lowerCAmelCase : str = True __lowerCAmelCase : Union[str, Any] = [element for element in array[i:] if element >= array[i]] __lowerCAmelCase : List[str] = longest_subsequence(lowercase__ ) if len(lowercase__ ) > len(lowercase__ ): __lowerCAmelCase : Optional[int] = temp_array else: i += 1 __lowerCAmelCase : Union[str, Any] = [element for element in array[1:] if element >= pivot] __lowerCAmelCase : List[str] = [pivot, *longest_subsequence(lowercase__ )] if len(lowercase__ ) > len(lowercase__ ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
275
def _lowercase ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __lowerCAmelCase : int = str(bin(lowercase__ ) )[2:] # remove the leading "0b" __lowerCAmelCase : Any = str(bin(lowercase__ ) )[2:] __lowerCAmelCase : List[str] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('''1''' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
275
1
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _lowercase ( ): with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(lowercase__ ): requests.request('''GET''' , '''https://huggingface.co''' ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request('''GET''' , '''https://huggingface.co''' , timeout=1.0 ) @pytest.mark.integration def _lowercase ( ): with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request('''GET''' , '''https://huggingface.co''' ) def _lowercase ( ): with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(lowercase__ ): http_head('''https://huggingface.co''' )
275
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
1
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->None: '''simple docstring''' warnings.warn( '''The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use BeitImageProcessor instead.''' , A_ , ) super().__init__(*A_ , **A_ )
275
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCAmelCase : Dict = Vector() def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(A_ ) , '''(0,0,0,0,0,1)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(A_ ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Vector([1, 2] ) __lowerCAmelCase : Optional[int] = Vector([1, 2, 3, 4, 5] ) __lowerCAmelCase : Optional[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCAmelCase : str = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : List[Any] = Vector([2, -1, 4] ) # for test of dot product __lowerCAmelCase : Optional[int] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : Any = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , A_ , A_ ) ) , '''(3,4,7)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) __lowerCAmelCase : Optional[Any] = x.copy() self.assertEqual(str(A_ ) , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(A_ ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Dict = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
275
1
def _lowercase ( lowercase__ ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowercase ( lowercase__ ): __lowerCAmelCase : Any = 0 __lowerCAmelCase : Union[str, Any] = number while duplicate > 0: __lowerCAmelCase, __lowerCAmelCase : Any = divmod(lowercase__ , 1_0 ) fact_sum += factorial(lowercase__ ) return fact_sum == number if __name__ == "__main__": print("Program to check whether a number is a Krisnamurthy Number or not.") _UpperCamelCase = int(input("Enter number: ").strip()) print( F"{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number." )
275
def _lowercase ( lowercase__ , lowercase__ ): if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
275
1
import logging from transformers import PretrainedConfig _UpperCamelCase = logging.getLogger(__name__) _UpperCamelCase = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """bertabs""" def __init__( self , A_=3_0522 , A_=512 , A_=6 , A_=512 , A_=8 , A_=512 , A_=0.2 , A_=6 , A_=768 , A_=8 , A_=2048 , A_=0.2 , **A_ , ) ->List[Any]: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : Optional[Any] = max_pos __lowerCAmelCase : Tuple = enc_layers __lowerCAmelCase : int = enc_hidden_size __lowerCAmelCase : Optional[int] = enc_heads __lowerCAmelCase : List[str] = enc_ff_size __lowerCAmelCase : Dict = enc_dropout __lowerCAmelCase : List[str] = dec_layers __lowerCAmelCase : Union[str, Any] = dec_hidden_size __lowerCAmelCase : Optional[int] = dec_heads __lowerCAmelCase : Tuple = dec_ff_size __lowerCAmelCase : Any = dec_dropout
275
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
1
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1
import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node _UpperCamelCase = 4 _UpperCamelCase = 3 class __lowercase (_UpperCAmelCase ): pass def _lowercase ( lowercase__ ): for shard in shards: for i in range(lowercase__ ): yield {"i": i, "shard": shard} def _lowercase ( ): __lowerCAmelCase : Any = int(os.environ['''RANK'''] ) __lowerCAmelCase : Union[str, Any] = int(os.environ['''WORLD_SIZE'''] ) __lowerCAmelCase : Tuple = ArgumentParser() parser.add_argument('''--streaming''' , type=lowercase__ ) parser.add_argument('''--local_rank''' , type=lowercase__ ) parser.add_argument('''--num_workers''' , type=lowercase__ , default=0 ) __lowerCAmelCase : List[Any] = parser.parse_args() __lowerCAmelCase : List[Any] = args.streaming __lowerCAmelCase : Dict = args.num_workers __lowerCAmelCase : Tuple = {'''shards''': [f"""shard_{shard_idx}""" for shard_idx in range(lowercase__ )]} __lowerCAmelCase : List[Any] = IterableDataset.from_generator(lowercase__ , gen_kwargs=lowercase__ ) if not streaming: __lowerCAmelCase : Dict = Dataset.from_list(list(lowercase__ ) ) __lowerCAmelCase : Optional[int] = split_dataset_by_node(lowercase__ , rank=lowercase__ , world_size=lowercase__ ) __lowerCAmelCase : Tuple = torch.utils.data.DataLoader(lowercase__ , num_workers=lowercase__ ) __lowerCAmelCase : List[Any] = NUM_SHARDS * NUM_ITEMS_PER_SHARD __lowerCAmelCase : Dict = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) __lowerCAmelCase : str = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(f"""local_size {local_size} != expected_local_size {expected_local_size}""" ) if __name__ == "__main__": main()
275
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ConsistencyModelPipeline _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt _UpperCamelCase = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[Any] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[str] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def UpperCamelCase__ ( self , A_=False ) ->Dict: '''simple docstring''' if class_cond: __lowerCAmelCase : List[str] = self.dummy_cond_unet else: __lowerCAmelCase : Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->Tuple: '''simple docstring''' if str(A_ ).startswith('''mps''' ): __lowerCAmelCase : str = torch.manual_seed(A_ ) else: __lowerCAmelCase : Dict = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Tuple = self.get_dummy_components() __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : str = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : str = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : List[Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : int = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : Any = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components(class_cond=A_ ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(**A_ ) __lowerCAmelCase : Union[str, Any] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_dummy_inputs(A_ ) __lowerCAmelCase : List[str] = 1 __lowerCAmelCase : Dict = None __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , A_=0 , A_=False , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->str: '''simple docstring''' __lowerCAmelCase : Dict = torch.manual_seed(A_ ) __lowerCAmelCase : Tuple = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __lowerCAmelCase : List[str] = self.get_fixed_latents(seed=A_ , device=A_ , dtype=A_ , shape=A_ ) __lowerCAmelCase : Union[str, Any] = latents return inputs def UpperCamelCase__ ( self , A_=0 , A_="cpu" , A_=torch.floataa , A_=(1, 3, 64, 64) ) ->Optional[int]: '''simple docstring''' if type(A_ ) == str: __lowerCAmelCase : int = torch.device(A_ ) __lowerCAmelCase : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) __lowerCAmelCase : Union[str, Any] = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) return latents def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : str = self.get_inputs() __lowerCAmelCase : Any = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.0_888, 0.0_881, 0.0_666, 0.0_479, 0.0_292, 0.0_195, 0.0_201, 0.0_163, 0.0_254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : List[str] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : List[Any] = self.get_inputs() __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : str = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] __lowerCAmelCase : List[Any] = np.array([0.0_340, 0.0_152, 0.0_063, 0.0_267, 0.0_221, 0.0_107, 0.0_416, 0.0_186, 0.0_217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Any = self.get_inputs(get_fixed_latents=A_ , device=A_ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : Dict = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[int] = np.array([0.1_875, 0.1_428, 0.1_289, 0.2_151, 0.2_092, 0.1_477, 0.1_877, 0.1_641, 0.1_353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __lowerCAmelCase : Union[str, Any] = ConsistencyModelPipeline(unet=A_ , scheduler=A_ ) pipe.to(torch_device=A_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Union[str, Any] = self.get_inputs(get_fixed_latents=A_ , device=A_ ) __lowerCAmelCase : Any = 1 __lowerCAmelCase : int = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A_ , enable_math=A_ , enable_mem_efficient=A_ ): __lowerCAmelCase : int = pipe(**A_ ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : str = image[0, -3:, -3:, -1] __lowerCAmelCase : Any = np.array([0.1_663, 0.1_948, 0.2_275, 0.1_680, 0.1_204, 0.1_245, 0.1_858, 0.1_338, 0.2_095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
275
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase = { "configuration_lilt": ["LILT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LiltConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ "LILT_PRETRAINED_MODEL_ARCHIVE_LIST", "LiltForQuestionAnswering", "LiltForSequenceClassification", "LiltForTokenClassification", "LiltModel", "LiltPreTrainedModel", ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
from collections import deque from .hash_table import HashTable class __lowercase (_UpperCAmelCase ): def __init__( self , *A_ , **A_ ) ->int: '''simple docstring''' super().__init__(*A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(A_ ) __lowerCAmelCase : int = self.values[key] def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return ( sum(self.charge_factor - len(A_ ) for slot in self.values ) / self.size_table * self.charge_factor ) def UpperCamelCase__ ( self , A_ , A_=None ) ->str: '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(A_ ) == 0 ): return key return super()._collision_resolution(A_ , A_ )
275
1
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase = "▁" _UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = BigBirdTokenizer _UpperCamelCase = BigBirdTokenizerFast _UpperCamelCase = True _UpperCamelCase = True def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' super().setUp() __lowerCAmelCase : Union[str, Any] = self.tokenizer_class(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''<s>''' __lowerCAmelCase : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ) , A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''[MASK]''' ) self.assertEqual(len(A_ ) , 1004 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' if not self.test_rust_tokenizer: return __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : Tuple = self.get_rust_tokenizer() __lowerCAmelCase : str = '''I was born in 92000, and this is falsé.''' __lowerCAmelCase : Union[str, Any] = tokenizer.tokenize(A_ ) __lowerCAmelCase : List[str] = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : Tuple = tokenizer.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : Dict = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : Union[str, Any] = self.get_rust_tokenizer() __lowerCAmelCase : Tuple = tokenizer.encode(A_ ) __lowerCAmelCase : Dict = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Dict = BigBirdTokenizer(A_ , keep_accents=A_ ) __lowerCAmelCase : int = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [285, 46, 10, 170, 382] , ) __lowerCAmelCase : List[str] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCAmelCase : int = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __lowerCAmelCase : Optional[Any] = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' return BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) @slow def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = '''Hello World!''' __lowerCAmelCase : Optional[Any] = [65, 1_8536, 2260, 101, 66] self.assertListEqual(A_ , self.big_tokenizer.encode(A_ ) ) @slow def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) # fmt: off __lowerCAmelCase : List[str] = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 3_4324, 497, 391, 408, 1_1342, 1244, 385, 100, 938, 985, 456, 574, 362, 1_2597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(A_ , self.big_tokenizer.encode(A_ ) ) @require_torch @slow def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __lowerCAmelCase : Optional[int] = list(self.big_tokenizer.get_vocab().keys() )[:10] __lowerCAmelCase : Tuple = ''' '''.join(A_ ) __lowerCAmelCase : str = self.big_tokenizer.encode_plus(A_ , return_tensors='''pt''' , return_token_type_ids=A_ ) __lowerCAmelCase : Optional[int] = self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=A_ ) __lowerCAmelCase : Tuple = BigBirdConfig(attention_type='''original_full''' ) __lowerCAmelCase : List[Any] = BigBirdModel(A_ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**A_ ) model(**A_ ) @slow def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) __lowerCAmelCase : Optional[Any] = tokenizer.decode(tokenizer('''Paris is the [MASK].''' ).input_ids ) self.assertTrue(decoded_text == '''[CLS] Paris is the[MASK].[SEP]''' ) @slow def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = {'''input_ids''': [[65, 3_9286, 458, 3_6335, 2001, 456, 1_3073, 1_3266, 455, 113, 7746, 1741, 1_1157, 391, 1_3073, 1_3266, 455, 113, 3967, 3_5412, 113, 4936, 109, 3870, 2377, 113, 3_0084, 4_5720, 458, 134, 1_7496, 112, 503, 1_1672, 113, 118, 112, 5665, 1_3347, 3_8687, 112, 1496, 3_1389, 112, 3268, 4_7264, 134, 962, 112, 1_6377, 8035, 2_3130, 430, 1_2169, 1_5518, 2_8592, 458, 146, 4_1697, 109, 391, 1_2169, 1_5518, 1_6689, 458, 146, 4_1358, 109, 452, 726, 4034, 111, 763, 3_5412, 5082, 388, 1903, 111, 9051, 391, 2870, 4_8918, 1900, 1123, 550, 998, 112, 9586, 1_5985, 455, 391, 410, 2_2955, 3_7636, 114, 66], [65, 448, 1_7496, 419, 3663, 385, 763, 113, 2_7533, 2870, 3283, 1_3043, 1639, 2_4713, 523, 656, 2_4013, 1_8550, 2521, 517, 2_7014, 2_1244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 1_1786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 2_1932, 1_8146, 726, 363, 1_7032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A_ , model_name='''google/bigbird-roberta-base''' , revision='''215c99f1600e06f83acce68422f2035b2b5c3510''' , )
275
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Optional[Any] = global_rng __lowerCAmelCase : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=1 , A_=0.0 , A_=1_6000 , A_=True , A_=True , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Any = min_seq_length __lowerCAmelCase : Tuple = max_seq_length __lowerCAmelCase : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Dict = feature_size __lowerCAmelCase : Optional[int] = padding_value __lowerCAmelCase : Tuple = sampling_rate __lowerCAmelCase : Union[str, Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Union[str, Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WavaVecaFeatureExtractor def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase__ ( self , A_ ) ->Optional[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Any = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Dict = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : List[Any] = np.asarray(A_ ) __lowerCAmelCase : Any = feat_extract(A_ , return_tensors='''np''' ).input_values __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : str = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : str = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Optional[int] = feat_extract(A_ , padding=A_ , max_length=A_ , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = range(800 , 1400 , 200 ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase : int = ['''longest''', '''max_length''', '''do_not_pad'''] __lowerCAmelCase : List[str] = [None, 1600, None] for max_length, padding in zip(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = feat_extract(A_ , max_length=A_ , padding=A_ ) __lowerCAmelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : List[str] = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) __lowerCAmelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : int = feat_extract( A_ , truncation=A_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Optional[int] = feat_extract( A_ , truncation=A_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) __lowerCAmelCase : List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' import torch __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Any = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase : List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase : Any = WavaVecaConfig.from_pretrained(A_ ) __lowerCAmelCase : Tuple = WavaVecaFeatureExtractor.from_pretrained(A_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
275
1
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = BertTokenizer _UpperCamelCase = BertTokenizerFast _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = filter_non_english def UpperCamelCase__ ( self ) ->int: '''simple docstring''' super().setUp() __lowerCAmelCase : Any = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCAmelCase : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = '''UNwant\u00E9d,running''' __lowerCAmelCase : Dict = '''unwanted, running''' return input_text, output_text def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.tokenizer_class(self.vocab_file ) __lowerCAmelCase : Any = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(A_ , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [9, 6, 7, 12, 10, 11] ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' if not self.test_rust_tokenizer: return __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : Tuple = self.get_rust_tokenizer() __lowerCAmelCase : List[Any] = '''UNwant\u00E9d,running''' __lowerCAmelCase : Dict = tokenizer.tokenize(A_ ) __lowerCAmelCase : Union[str, Any] = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : int = tokenizer.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : Optional[Any] = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : Union[str, Any] = self.get_rust_tokenizer() __lowerCAmelCase : Tuple = tokenizer.encode(A_ ) __lowerCAmelCase : Optional[int] = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # With lower casing __lowerCAmelCase : Optional[int] = self.get_tokenizer(do_lower_case=A_ ) __lowerCAmelCase : Tuple = self.get_rust_tokenizer(do_lower_case=A_ ) __lowerCAmelCase : Union[str, Any] = '''UNwant\u00E9d,running''' __lowerCAmelCase : Union[str, Any] = tokenizer.tokenize(A_ ) __lowerCAmelCase : Optional[Any] = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : Tuple = tokenizer.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : List[str] = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : Tuple = self.get_rust_tokenizer() __lowerCAmelCase : Optional[Any] = tokenizer.encode(A_ ) __lowerCAmelCase : Union[str, Any] = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Optional[Any] = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Any = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Any = BasicTokenizer(do_lower_case=A_ , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = BasicTokenizer() __lowerCAmelCase : List[str] = '''a\n\'ll !!to?\'d of, can\'t.''' __lowerCAmelCase : Dict = ['''a''', '''\'''', '''ll''', '''!''', '''!''', '''to''', '''?''', '''\'''', '''d''', '''of''', ''',''', '''can''', '''\'''', '''t''', '''.'''] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] __lowerCAmelCase : List[str] = {} for i, token in enumerate(A_ ): __lowerCAmelCase : Dict = i __lowerCAmelCase : List[Any] = WordpieceTokenizer(vocab=A_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = self.get_tokenizer() __lowerCAmelCase : str = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A_ ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) self.assertListEqual( [rust_tokenizer.tokenize(A_ ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[Any] = self.tokenizer_class.from_pretrained('''bert-base-uncased''' ) __lowerCAmelCase : Tuple = tokenizer.encode('''sequence builders''' , add_special_tokens=A_ ) __lowerCAmelCase : Any = tokenizer.encode('''multi-sequence build''' , add_special_tokens=A_ ) __lowerCAmelCase : Optional[Any] = tokenizer.build_inputs_with_special_tokens(A_ ) __lowerCAmelCase : Tuple = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Optional[int] = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" __lowerCAmelCase : Dict = tokenizer_r.encode_plus( A_ , return_attention_mask=A_ , return_token_type_ids=A_ , return_offsets_mapping=A_ , add_special_tokens=A_ , ) __lowerCAmelCase : Optional[int] = tokenizer_r.do_lower_case if hasattr(A_ , '''do_lower_case''' ) else False __lowerCAmelCase : Union[str, Any] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''A'''), ((1, 2), ''','''), ((3, 5), '''na'''), ((5, 6), '''##ï'''), ((6, 8), '''##ve'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''Allen'''), ((21, 23), '''##NL'''), ((23, 24), '''##P'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''a'''), ((1, 2), ''','''), ((3, 8), '''naive'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''allen'''), ((21, 23), '''##nl'''), ((23, 24), '''##p'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['''offset_mapping'''] ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = ['''的''', '''人''', '''有'''] __lowerCAmelCase : Tuple = ''''''.join(A_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : List[str] = self.tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Optional[Any] = tokenizer_p.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : Optional[int] = tokenizer_r.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(A_ ) __lowerCAmelCase : List[str] = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) __lowerCAmelCase : int = False __lowerCAmelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Optional[int] = self.tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCAmelCase : Dict = tokenizer_r.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : str = tokenizer_p.encode(A_ , add_special_tokens=A_ ) __lowerCAmelCase : List[str] = tokenizer_r.convert_ids_to_tokens(A_ ) __lowerCAmelCase : List[str] = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that only the first Chinese character is not preceded by "##". __lowerCAmelCase : Any = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(A_ ) ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ )
275
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=False , A_=True , A_="None" , A_=3 , A_=4 , A_=None , ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = parent __lowerCAmelCase : List[str] = batch_size __lowerCAmelCase : Dict = seq_length __lowerCAmelCase : List[Any] = is_training __lowerCAmelCase : List[Any] = use_input_mask __lowerCAmelCase : Optional[int] = use_token_type_ids __lowerCAmelCase : Tuple = use_labels __lowerCAmelCase : str = vocab_size __lowerCAmelCase : int = hidden_size __lowerCAmelCase : Any = num_hidden_layers __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : int = hidden_act __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Union[str, Any] = type_sequence_label_size __lowerCAmelCase : Optional[int] = initializer_range __lowerCAmelCase : int = num_labels __lowerCAmelCase : int = num_choices __lowerCAmelCase : List[str] = relative_attention __lowerCAmelCase : Union[str, Any] = position_biased_input __lowerCAmelCase : int = pos_att_type __lowerCAmelCase : List[Any] = scope def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : int = None if self.use_input_mask: __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __lowerCAmelCase : List[str] = None if self.use_token_type_ids: __lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : int = None __lowerCAmelCase : List[str] = None if self.use_labels: __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_config() __lowerCAmelCase : Dict = 300 return config def UpperCamelCase__ ( self , A_ ) ->Union[str, Any]: '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = DebertaModel(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ )[0] __lowerCAmelCase : Any = model(A_ , token_type_ids=A_ )[0] __lowerCAmelCase : List[str] = model(A_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = DebertaForMaskedLM(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = self.num_labels __lowerCAmelCase : Tuple = DebertaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->Dict: '''simple docstring''' __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Optional[int] = DebertaForTokenClassification(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Tuple = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : List[str] = DebertaForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : int = model( A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ( __lowerCAmelCase ), ) : Tuple = config_and_inputs __lowerCAmelCase : Tuple = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowercase (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : int = DebertaModelTester(self ) __lowerCAmelCase : List[Any] = ConfigTester(self , config_class=A_ , hidden_size=37 ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*A_ ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*A_ ) @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[int] = DebertaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowercase (unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' pass @slow def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = DebertaModel.from_pretrained('''microsoft/deberta-base''' ) __lowerCAmelCase : Tuple = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase : Optional[int] = model(A_ , attention_mask=A_ )[0] # compare the actual values for a slice. __lowerCAmelCase : Optional[Any] = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , A_ , atol=1e-4 ) , f"""{output[:, 1:4, 1:4]}""" )
275
1
def _lowercase ( lowercase__ ): return "".join(chr(ord(lowercase__ ) - 3_2 ) if '''a''' <= char <= '''z''' else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
275
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowercase__ ): __lowerCAmelCase : str = [] __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : str = [] for rt in rc.restypes: __lowerCAmelCase : List[Any] = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase : List[str] = {name: i for i, name in enumerate(lowercase__ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) __lowerCAmelCase : List[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Optional[Any] = torch.tensor( lowercase__ , dtype=torch.intaa , device=protein['''aatype'''].device , ) __lowerCAmelCase : Tuple = torch.tensor( lowercase__ , dtype=torch.floataa , device=protein['''aatype'''].device , ) __lowerCAmelCase : List[Any] = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase : Any = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : int = residx_atomaa_mask __lowerCAmelCase : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase : int = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase : Union[str, Any] = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase : str = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase : Optional[int] = rc.restype_atoa[restype_letter] __lowerCAmelCase : Optional[Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase : str = rc.atom_order[atom_name] __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = restype_atomaa_mask[protein_aatype] __lowerCAmelCase : Any = residx_atomaa_mask return protein def _lowercase ( lowercase__ ): __lowerCAmelCase : Dict = tree_map(lambda lowercase__ : torch.tensor(lowercase__ , device=batch['''aatype'''].device ) , lowercase__ , np.ndarray ) __lowerCAmelCase : Tuple = tensor_tree_map(lambda lowercase__ : np.array(lowercase__ ) , make_atomaa_masks(lowercase__ ) ) return out
275
1
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, "constant": get_constant_schedule, "constant_w_warmup": get_constant_schedule_with_warmup, } class __lowercase (_UpperCAmelCase ): def __init__( self , A_=None , A_=None , *A_ , **A_ ) ->Optional[Any]: '''simple docstring''' super().__init__(*A_ , **A_ ) if config is None: assert isinstance(self.model , A_ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f""" {self.model.__class__}""" ) __lowerCAmelCase : int = self.model.config else: __lowerCAmelCase : Tuple = config __lowerCAmelCase : Union[str, Any] = data_args __lowerCAmelCase : Optional[Any] = self.config.tgt_vocab_size if isinstance(self.config , A_ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: __lowerCAmelCase : Tuple = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss __lowerCAmelCase : int = label_smoothed_nll_loss def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' if self.optimizer is None: __lowerCAmelCase : Dict = ['''bias''', '''LayerNorm.weight'''] __lowerCAmelCase : Optional[Any] = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] __lowerCAmelCase : str = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: __lowerCAmelCase : Optional[Any] = Adafactor __lowerCAmelCase : Tuple = {'''scale_parameter''': False, '''relative_step''': False} else: __lowerCAmelCase : Any = AdamW __lowerCAmelCase : int = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } __lowerCAmelCase : Any = self.args.learning_rate if self.sharded_ddp: __lowerCAmelCase : Tuple = OSS( params=A_ , optim=A_ , **A_ , ) else: __lowerCAmelCase : Optional[int] = optimizer_cls(A_ , **A_ ) if self.lr_scheduler is None: __lowerCAmelCase : List[str] = self._get_lr_scheduler(A_ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase__ ( self , A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Dict = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": __lowerCAmelCase : Tuple = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": __lowerCAmelCase : Optional[int] = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: __lowerCAmelCase : List[str] = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A_ ) return scheduler def UpperCamelCase__ ( self ) ->Optional[torch.utils.data.Sampler]: '''simple docstring''' if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->str: '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token __lowerCAmelCase : Tuple = model(**A_ , use_cache=A_ )[0] __lowerCAmelCase : Optional[int] = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models __lowerCAmelCase, __lowerCAmelCase : Tuple = model(**A_ , labels=A_ , use_cache=A_ )[:2] else: # compute label smoothed loss __lowerCAmelCase : Tuple = model(**A_ , use_cache=A_ )[0] __lowerCAmelCase : Optional[int] = torch.nn.functional.log_softmax(A_ , dim=-1 ) __lowerCAmelCase, __lowerCAmelCase : List[Any] = self.loss_fn(A_ , A_ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase__ ( self , A_ , A_ ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : int = inputs.pop('''labels''' ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self._compute_loss(A_ , A_ , A_ ) return loss def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , ) ->Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: '''simple docstring''' __lowerCAmelCase : int = self._prepare_inputs(A_ ) __lowerCAmelCase : int = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: __lowerCAmelCase : List[Any] = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A_ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: __lowerCAmelCase : Optional[Any] = self._pad_tensors_to_max_len(A_ , gen_kwargs['''max_length'''] ) __lowerCAmelCase : int = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data __lowerCAmelCase, __lowerCAmelCase : List[str] = self._compute_loss(A_ , A_ , A_ ) __lowerCAmelCase : Union[str, Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) __lowerCAmelCase : Union[str, Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: __lowerCAmelCase : Any = self._pad_tensors_to_max_len(A_ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase__ ( self , A_ , A_ ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' f""" padded to `max_length`={max_length}""" ) __lowerCAmelCase : Optional[int] = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) __lowerCAmelCase : List[str] = tensor return padded_tensor
275
def _lowercase ( lowercase__ ): if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) __lowerCAmelCase : int = sorted(string.lower() ) return len(lowercase__ ) == len(set(lowercase__ ) ) if __name__ == "__main__": _UpperCamelCase = input("Enter a string ").strip() _UpperCamelCase = is_isogram(input_str) print(F"{input_str} is {'an' if isogram else 'not an'} isogram.")
275
1
from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP _UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name _UpperCamelCase = "\n Examples:\n ```py\n >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\")\n >>> pipe_prior.to(\"cuda\")\n\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> negative_image_emb = out.negative_image_embeds\n\n >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\")\n >>> pipe.to(\"cuda\")\n\n >>> image = pipe(\n ... prompt,\n ... image_embeds=image_emb,\n ... negative_image_embeds=negative_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... ).images\n\n >>> image[0].save(\"cat.png\")\n ```\n" def _lowercase ( lowercase__ , lowercase__ , lowercase__=8 ): __lowerCAmelCase : Tuple = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 __lowerCAmelCase : List[str] = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class __lowercase (_UpperCAmelCase ): def __init__( self , A_ , A_ , A_ , A_ , A_ , ) ->str: '''simple docstring''' super().__init__() self.register_modules( text_encoder=A_ , tokenizer=A_ , unet=A_ , scheduler=A_ , movq=A_ , ) __lowerCAmelCase : str = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ ) ->List[Any]: '''simple docstring''' if latents is None: __lowerCAmelCase : int = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) __lowerCAmelCase : List[Any] = latents.to(A_ ) __lowerCAmelCase : Any = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_=None , ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[str] = len(A_ ) if isinstance(A_ , A_ ) else 1 # get prompt text embeddings __lowerCAmelCase : str = self.tokenizer( A_ , padding='''max_length''' , truncation=A_ , max_length=77 , return_attention_mask=A_ , add_special_tokens=A_ , return_tensors='''pt''' , ) __lowerCAmelCase : Optional[Any] = text_inputs.input_ids __lowerCAmelCase : Dict = self.tokenizer(A_ , padding='''longest''' , return_tensors='''pt''' ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(A_ , A_ ): __lowerCAmelCase : str = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) __lowerCAmelCase : List[Any] = text_input_ids.to(A_ ) __lowerCAmelCase : Tuple = text_inputs.attention_mask.to(A_ ) __lowerCAmelCase, __lowerCAmelCase : int = self.text_encoder( input_ids=A_ , attention_mask=A_ ) __lowerCAmelCase : Dict = prompt_embeds.repeat_interleave(A_ , dim=0 ) __lowerCAmelCase : str = text_encoder_hidden_states.repeat_interleave(A_ , dim=0 ) __lowerCAmelCase : Dict = text_mask.repeat_interleave(A_ , dim=0 ) if do_classifier_free_guidance: __lowerCAmelCase : List[str] if negative_prompt is None: __lowerCAmelCase : int = [''''''] * batch_size elif type(A_ ) is not type(A_ ): raise TypeError( f"""`negative_prompt` should be the same type to `prompt`, but got {type(A_ )} !=""" f""" {type(A_ )}.""" ) elif isinstance(A_ , A_ ): __lowerCAmelCase : Tuple = [negative_prompt] elif batch_size != len(A_ ): raise ValueError( f"""`negative_prompt`: {negative_prompt} has batch size {len(A_ )}, but `prompt`:""" f""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" ''' the batch size of `prompt`.''' ) else: __lowerCAmelCase : str = negative_prompt __lowerCAmelCase : int = self.tokenizer( A_ , padding='''max_length''' , max_length=77 , truncation=A_ , return_attention_mask=A_ , add_special_tokens=A_ , return_tensors='''pt''' , ) __lowerCAmelCase : int = uncond_input.input_ids.to(A_ ) __lowerCAmelCase : Union[str, Any] = uncond_input.attention_mask.to(A_ ) __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = self.text_encoder( input_ids=A_ , attention_mask=A_ ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __lowerCAmelCase : List[Any] = negative_prompt_embeds.shape[1] __lowerCAmelCase : str = negative_prompt_embeds.repeat(1 , A_ ) __lowerCAmelCase : Union[str, Any] = negative_prompt_embeds.view(batch_size * num_images_per_prompt , A_ ) __lowerCAmelCase : Optional[Any] = uncond_text_encoder_hidden_states.shape[1] __lowerCAmelCase : Union[str, Any] = uncond_text_encoder_hidden_states.repeat(1 , A_ , 1 ) __lowerCAmelCase : Optional[int] = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , A_ , -1 ) __lowerCAmelCase : Union[str, Any] = uncond_text_mask.repeat_interleave(A_ , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __lowerCAmelCase : Union[str, Any] = torch.cat([negative_prompt_embeds, prompt_embeds] ) __lowerCAmelCase : str = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) __lowerCAmelCase : Optional[int] = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def UpperCamelCase__ ( self , A_=0 ) ->int: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) __lowerCAmelCase : Optional[Any] = torch.device(f"""cuda:{gpu_id}""" ) __lowerCAmelCase : Dict = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def UpperCamelCase__ ( self , A_=0 ) ->Optional[int]: '''simple docstring''' if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) __lowerCAmelCase : Tuple = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __lowerCAmelCase : Union[str, Any] = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: __lowerCAmelCase, __lowerCAmelCase : Tuple = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) if self.safety_checker is not None: __lowerCAmelCase, __lowerCAmelCase : List[Any] = cpu_offload_with_hook(self.safety_checker , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. __lowerCAmelCase : Optional[int] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ) ->str: '''simple docstring''' if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self , A_ , A_ , A_ , A_ = None , A_ = 512 , A_ = 512 , A_ = 100 , A_ = 4.0 , A_ = 1 , A_ = None , A_ = None , A_ = "pil" , A_ = True , ) ->Dict: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : List[str] = 1 elif isinstance(A_ , A_ ): __lowerCAmelCase : Union[str, Any] = len(A_ ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(A_ )}""" ) __lowerCAmelCase : Union[str, Any] = self._execution_device __lowerCAmelCase : Any = batch_size * num_images_per_prompt __lowerCAmelCase : Tuple = guidance_scale > 1.0 __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[Any] = self._encode_prompt( A_ , A_ , A_ , A_ , A_ ) if isinstance(A_ , A_ ): __lowerCAmelCase : Dict = torch.cat(A_ , dim=0 ) if isinstance(A_ , A_ ): __lowerCAmelCase : Dict = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: __lowerCAmelCase : Dict = image_embeds.repeat_interleave(A_ , dim=0 ) __lowerCAmelCase : Optional[Any] = negative_image_embeds.repeat_interleave(A_ , dim=0 ) __lowerCAmelCase : Optional[int] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) __lowerCAmelCase : Any = self.scheduler.timesteps __lowerCAmelCase : Union[str, Any] = self.unet.config.in_channels __lowerCAmelCase, __lowerCAmelCase : Optional[int] = get_new_h_w(A_ , A_ , self.movq_scale_factor ) # create initial latent __lowerCAmelCase : Optional[Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance __lowerCAmelCase : str = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __lowerCAmelCase : List[str] = {'''text_embeds''': prompt_embeds, '''image_embeds''': image_embeds} __lowerCAmelCase : Any = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: __lowerCAmelCase, __lowerCAmelCase : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = noise_pred.chunk(2 ) __lowerCAmelCase, __lowerCAmelCase : str = variance_pred.chunk(2 ) __lowerCAmelCase : Tuple = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __lowerCAmelCase : Optional[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __lowerCAmelCase : int = self.scheduler.step( A_ , A_ , A_ , generator=A_ , ).prev_sample # post-processing __lowerCAmelCase : List[str] = self.movq.decode(A_ , force_not_quantize=A_ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __lowerCAmelCase : Dict = image * 0.5 + 0.5 __lowerCAmelCase : int = image.clamp(0 , 1 ) __lowerCAmelCase : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __lowerCAmelCase : Dict = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
275
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
1
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _UpperCamelCase = 16 _UpperCamelCase = 32 def _lowercase ( lowercase__ , lowercase__ = 1_6 ): __lowerCAmelCase : List[str] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) __lowerCAmelCase : Tuple = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) __lowerCAmelCase : List[str] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __lowerCAmelCase : Optional[int] = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __lowerCAmelCase : str = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. __lowerCAmelCase : List[Any] = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __lowerCAmelCase : Union[str, Any] = 1_6 elif accelerator.mixed_precision != "no": __lowerCAmelCase : str = 8 else: __lowerCAmelCase : Optional[Any] = None return tokenizer.pad( lowercase__ , padding='''longest''' , max_length=lowercase__ , pad_to_multiple_of=lowercase__ , return_tensors='''pt''' , ) # Instantiate dataloaders. __lowerCAmelCase : Optional[int] = DataLoader( tokenized_datasets['''train'''] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ , drop_last=lowercase__ ) __lowerCAmelCase : int = DataLoader( tokenized_datasets['''validation'''] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ , drop_last=(accelerator.mixed_precision == '''fp8''') , ) return train_dataloader, eval_dataloader def _lowercase ( lowercase__ , lowercase__ ): # Initialize accelerator __lowerCAmelCase : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __lowerCAmelCase : Optional[int] = config['''lr'''] __lowerCAmelCase : List[str] = int(config['''num_epochs'''] ) __lowerCAmelCase : Any = int(config['''seed'''] ) __lowerCAmelCase : Optional[int] = int(config['''batch_size'''] ) __lowerCAmelCase : List[Any] = evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation __lowerCAmelCase : str = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: __lowerCAmelCase : Dict = batch_size // MAX_GPU_BATCH_SIZE __lowerCAmelCase : Tuple = MAX_GPU_BATCH_SIZE set_seed(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_dataloaders(lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __lowerCAmelCase : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=lowercase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __lowerCAmelCase : Dict = model.to(accelerator.device ) # Instantiate optimizer __lowerCAmelCase : Dict = AdamW(params=model.parameters() , lr=lowercase__ ) # Instantiate scheduler __lowerCAmelCase : List[Any] = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=1_0_0 , num_training_steps=(len(lowercase__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # Now we train the model for epoch in range(lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __lowerCAmelCase : Union[str, Any] = model(**lowercase__ ) __lowerCAmelCase : Tuple = outputs.loss __lowerCAmelCase : str = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __lowerCAmelCase : int = model(**lowercase__ ) __lowerCAmelCase : Optional[int] = outputs.logits.argmax(dim=-1 ) __lowerCAmelCase, __lowerCAmelCase : Union[str, Any] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) __lowerCAmelCase : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , lowercase__ ) def _lowercase ( ): __lowerCAmelCase : Any = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=lowercase__ , default=lowercase__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) __lowerCAmelCase : Dict = parser.parse_args() __lowerCAmelCase : Optional[Any] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 4_2, '''batch_size''': 1_6} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
275
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = np.full((len(lowercase__ ), sequence_length, 2) , lowercase__ ) else: __lowerCAmelCase : Optional[int] = np.full((len(lowercase__ ), sequence_length) , lowercase__ ) for i, tensor in enumerate(lowercase__ ): if padding_side == "right": if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : int = tensor[:sequence_length] else: if isinstance(lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = tensor[:sequence_length] else: __lowerCAmelCase : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = ord(lowercase__ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True __lowerCAmelCase : int = unicodedata.category(lowercase__ ) if cat.startswith('''P''' ): return True return False @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = True _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = -100 _UpperCamelCase = "pt" def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' import torch __lowerCAmelCase : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCAmelCase : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCAmelCase : List[Any] = self.tokenizer.pad( A_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCAmelCase : Dict = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCAmelCase : Optional[int] = self.tokenizer.padding_side if padding_side == "right": __lowerCAmelCase : Any = [ list(A_ ) + [self.label_pad_token_id] * (sequence_length - len(A_ )) for label in labels ] else: __lowerCAmelCase : Optional[int] = [ [self.label_pad_token_id] * (sequence_length - len(A_ )) + list(A_ ) for label in labels ] __lowerCAmelCase : Tuple = [feature['''ner_tags'''] for feature in features] __lowerCAmelCase : List[Any] = padding_tensor(A_ , -1 , A_ , A_ ) __lowerCAmelCase : Optional[int] = [feature['''original_entity_spans'''] for feature in features] __lowerCAmelCase : Any = padding_tensor(A_ , (-1, -1) , A_ , A_ ) __lowerCAmelCase : Optional[Any] = {k: torch.tensor(A_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
275
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _lowercase ( lowercase__=None ): if subparsers is not None: __lowerCAmelCase : str = subparsers.add_parser('''test''' ) else: __lowerCAmelCase : int = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=lowercase__ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=lowercase__ ) return parser def _lowercase ( lowercase__ ): __lowerCAmelCase : Union[str, Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: __lowerCAmelCase : Dict = script_name else: __lowerCAmelCase : Union[str, Any] = f"""--config_file={args.config_file} {script_name}""" __lowerCAmelCase : Tuple = ['''accelerate-launch'''] + test_args.split() __lowerCAmelCase : Any = execute_subprocess_async(lowercase__ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def _lowercase ( ): __lowerCAmelCase : Optional[Any] = test_command_parser() __lowerCAmelCase : Optional[Any] = parser.parse_args() test_command(lowercase__ ) if __name__ == "__main__": main()
275
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase (unittest.TestCase ): _UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def UpperCamelCase__ ( self ) ->str: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : Tuple = (3, 32, 128) __lowerCAmelCase : List[str] = tempfile.mkdtemp() # fmt: off __lowerCAmelCase : List[str] = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on __lowerCAmelCase : Optional[int] = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) __lowerCAmelCase : Union[str, Any] = { '''do_normalize''': False, '''do_resize''': True, '''image_processor_type''': '''ViTImageProcessor''', '''resample''': 3, '''size''': {'''height''': 32, '''width''': 128}, } __lowerCAmelCase : Optional[Any] = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(A_ , A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Tuple: '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Tuple = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) __lowerCAmelCase : str = Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) return image_input def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Union[str, Any] = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = self.get_image_processor() __lowerCAmelCase : List[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : int = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) __lowerCAmelCase : int = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Any = self.get_image_processor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Optional[int] = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = image_processor(A_ , return_tensors='''np''' ) __lowerCAmelCase : Tuple = processor(images=A_ , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Union[str, Any] = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Any = '''test''' __lowerCAmelCase : Dict = processor(text=A_ ) __lowerCAmelCase : str = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : str = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = '''test''' __lowerCAmelCase : int = self.prepare_image_inputs() __lowerCAmelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''pixel_values''', '''labels'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = self.get_image_processor() __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Optional[int] = processor.char_decode(A_ ) __lowerCAmelCase : Tuple = tokenizer.batch_decode(A_ ) __lowerCAmelCase : Any = [seq.replace(''' ''' , '''''' ) for seq in decoded_tok] self.assertListEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : int = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = self.prepare_image_inputs() __lowerCAmelCase : List[Any] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_image_processor() __lowerCAmelCase : List[str] = self.get_tokenizer() __lowerCAmelCase : Any = MgpstrProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[Any] = torch.randn(1 , 27 , 38 ) __lowerCAmelCase : Optional[int] = torch.randn(1 , 27 , 5_0257 ) __lowerCAmelCase : Optional[Any] = torch.randn(1 , 27 , 3_0522 ) __lowerCAmelCase : List[str] = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['''generated_text''', '''scores''', '''char_preds''', '''bpe_preds''', '''wp_preds'''] )
275
1
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def _lowercase ( lowercase__ ): __lowerCAmelCase : Tuple = os.path.join(args.tf_model_dir , '''parameters.json''' ) __lowerCAmelCase : Tuple = json.loads(open(lowercase__ ).read() ) if not params: raise ValueError( f"""It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.""" ) if not args.output.endswith('''.pt''' ): __lowerCAmelCase : Tuple = args.output + '''.pt''' __lowerCAmelCase : int = OrderedDict() with tf.device('''/CPU:0''' ): __lowerCAmelCase : Optional[int] = tf.train.load_checkpoint(args.tf_model_dir ) __lowerCAmelCase : List[str] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): __lowerCAmelCase : str = reader.get_tensor(lowercase__ ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): __lowerCAmelCase : List[Any] = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): __lowerCAmelCase : List[Any] = 8 __lowerCAmelCase : List[str] = '''model.sqout.%d.weight''' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time __lowerCAmelCase : Any = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Union[str, Any] = torch.tensor(lowercase__ ) elif key_name.startswith('''model/moe''' ): __lowerCAmelCase : Union[str, Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): __lowerCAmelCase : Optional[Any] = '''model.blocks.%d.feed_forward.mlp.router.classifier.weight''' % player __lowerCAmelCase : Dict = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : int = torch.tensor(lowercase__ ) elif key_name.endswith('''/softmlp/kernel''' ): __lowerCAmelCase : Union[str, Any] = '''model.blocks.%d.feed_forward.soft_bypass_mlp.weight''' % player __lowerCAmelCase : int = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Any = torch.tensor(lowercase__ ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): __lowerCAmelCase : Union[str, Any] = key_name[-9:-7] for i in range(1_6 ): __lowerCAmelCase : Dict = '''model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight''' % (player, i, nlayer) __lowerCAmelCase : List[Any] = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided __lowerCAmelCase : Optional[Any] = torch.tensor(lowercase__ ) elif key_name.startswith('''model/mlp''' ): __lowerCAmelCase : List[Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): __lowerCAmelCase : Tuple = '''model.blocks.%d.feed_forward.mlp.wi.weight''' % player __lowerCAmelCase : List[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Union[str, Any] = torch.tensor(lowercase__ ) elif key_name.endswith('''/p1/bias''' ): __lowerCAmelCase : Union[str, Any] = '''model.blocks.%d.feed_forward.mlp.wi.bias''' % player __lowerCAmelCase : Dict = vnp.copy() # same because it is one dimensional __lowerCAmelCase : List[str] = torch.tensor(lowercase__ ) elif key_name.endswith('''/p2/kernel''' ): __lowerCAmelCase : str = '''model.blocks.%d.feed_forward.mlp.wo.weight''' % player __lowerCAmelCase : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Tuple = torch.tensor(lowercase__ ) elif key_name.endswith('''/p2/bias''' ): __lowerCAmelCase : int = '''model.blocks.%d.feed_forward.mlp.wo.bias''' % player __lowerCAmelCase : Union[str, Any] = vnp.copy() # same because it is one dimensional __lowerCAmelCase : str = torch.tensor(lowercase__ ) elif key_name.startswith('''model/ln''' ): __lowerCAmelCase : Union[str, Any] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): __lowerCAmelCase : List[Any] = '''model.blocks.%d.feed_forward.norm.bias''' % player __lowerCAmelCase : Optional[Any] = vnp.copy() # same because it is one dimensional __lowerCAmelCase : List[str] = torch.tensor(lowercase__ ) elif key_name.endswith('''/g''' ): __lowerCAmelCase : Union[str, Any] = '''model.blocks.%d.feed_forward.norm.weight''' % player __lowerCAmelCase : Union[str, Any] = vnp.copy() # same because it is one dimensional __lowerCAmelCase : Dict = torch.tensor(lowercase__ ) elif key_name.startswith('''model/att''' ): __lowerCAmelCase : List[Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): __lowerCAmelCase : Optional[int] = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum __lowerCAmelCase : Tuple = state[:, 0, :, :] __lowerCAmelCase : Optional[int] = state[:, 1, :, :] __lowerCAmelCase : List[str] = state[:, 2, :, :] __lowerCAmelCase : Optional[int] = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Dict = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Optional[int] = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Union[str, Any] = '''model.blocks.%d.self_attn.self_attn.q_proj.weight''' % player __lowerCAmelCase : List[str] = torch.tensor(lowercase__ ) __lowerCAmelCase : Dict = '''model.blocks.%d.self_attn.self_attn.k_proj.weight''' % player __lowerCAmelCase : List[Any] = torch.tensor(lowercase__ ) __lowerCAmelCase : Any = '''model.blocks.%d.self_attn.self_attn.v_proj.weight''' % player __lowerCAmelCase : str = torch.tensor(lowercase__ ) elif key_name.endswith('''/o/kernel''' ): __lowerCAmelCase : List[str] = '''model.blocks.%d.self_attn.self_attn.out_proj.weight''' % player __lowerCAmelCase : Optional[int] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Optional[int] = torch.tensor(lowercase__ ) elif key_name.startswith('''model/an''' ): __lowerCAmelCase : List[str] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): __lowerCAmelCase : Dict = '''model.blocks.%d.self_attn.norm.bias''' % player __lowerCAmelCase : str = vnp.copy() # same because it is one dimensional __lowerCAmelCase : str = torch.tensor(lowercase__ ) elif key_name.endswith('''/g''' ): __lowerCAmelCase : Optional[int] = '''model.blocks.%d.self_attn.norm.weight''' % player __lowerCAmelCase : int = vnp.copy() # same because it is one dimensional __lowerCAmelCase : List[str] = torch.tensor(lowercase__ ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): __lowerCAmelCase : int = {'''wte''': '''embed_tokens''', '''wpe''': '''position_embeddings''', '''ete''': '''extra_position_embeddings'''}[ key_name[-3:] ] __lowerCAmelCase : Dict = '''model.%s.weight''' % nlayer __lowerCAmelCase : List[str] = vnp.copy() # same in embedded __lowerCAmelCase : Optional[int] = torch.tensor(lowercase__ ) if key_name.startswith('''model/wte''' ): __lowerCAmelCase : Union[str, Any] = '''lm_head.weight''' __lowerCAmelCase : Dict = vnp.copy() # same in embedded __lowerCAmelCase : List[str] = torch.tensor(lowercase__ ) elif key_name.startswith('''model/wob''' ): __lowerCAmelCase : Dict = '''final_logits_bias''' __lowerCAmelCase : Tuple = vnp.copy() # same in embedded __lowerCAmelCase : str = state.reshape((1, -1) ) __lowerCAmelCase : Tuple = torch.tensor(lowercase__ ) elif key_name == "model/dense/kernel": __lowerCAmelCase : Optional[Any] = '''model.last_project.weight''' __lowerCAmelCase : Any = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix __lowerCAmelCase : Optional[int] = torch.tensor(lowercase__ ) elif key_name == "model/dense_1/bias": __lowerCAmelCase : Dict = '''model.last_project.bias''' __lowerCAmelCase : int = vnp.copy() # same because it is one dimensional __lowerCAmelCase : Optional[int] = torch.tensor(lowercase__ ) torch.save(lowercase__ , args.output ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser( description="model converter.", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument("--tf_model_dir", metavar="PATH", type=str, required=True, help="import model") parser.add_argument("--output", metavar="PATH", type=str, required=True, help="output model") _UpperCamelCase = parser.parse_args() convert_tf_gptsan_to_pt(args)
275
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowercase (unittest.TestCase ): @property def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = self.dummy_uncond_unet __lowerCAmelCase : Any = PNDMScheduler() __lowerCAmelCase : Dict = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' ).images __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[Any] = pndm(generator=A_ , num_inference_steps=20 , output_type='''numpy''' , return_dict=A_ )[0] __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = '''google/ddpm-cifar10-32''' __lowerCAmelCase : Union[str, Any] = UNetaDModel.from_pretrained(A_ ) __lowerCAmelCase : int = PNDMScheduler() __lowerCAmelCase : Any = PNDMPipeline(unet=A_ , scheduler=A_ ) pndm.to(A_ ) pndm.set_progress_bar_config(disable=A_ ) __lowerCAmelCase : Tuple = torch.manual_seed(0 ) __lowerCAmelCase : Any = pndm(generator=A_ , output_type='''numpy''' ).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
275
1
# Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib _UpperCamelCase = get_logger() _UpperCamelCase = None class __lowercase (TensorFormatter[Mapping, """jax.Array""", Mapping] ): def __init__( self , A_=None , A_=None , **A_ ) ->List[Any]: '''simple docstring''' super().__init__(features=A_ ) import jax from jaxlib.xla_client import Device if isinstance(A_ , A_ ): raise ValueError( f"""Expected {device} to be a `str` not {type(A_ )}, as `jaxlib.xla_extension.Device` """ '''is not serializable neither with `pickle` nor with `dill`. Instead you can surround ''' '''the device with `str()` to get its string identifier that will be internally mapped ''' '''to the actual `jaxlib.xla_extension.Device`.''' ) __lowerCAmelCase : Dict = device if isinstance(A_ , A_ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: __lowerCAmelCase : List[Any] = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( f"""Device with string identifier {self.device} not listed among the available """ f"""devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default """ f"""device: {str(jax.devices()[0] )}.""" ) __lowerCAmelCase : Optional[Any] = str(jax.devices()[0] ) __lowerCAmelCase : Dict = jnp_array_kwargs @staticmethod def UpperCamelCase__ ( ) ->Dict[str, "jaxlib.xla_extension.Device"]: '''simple docstring''' import jax return {str(A_ ): device for device in jax.devices()} def UpperCamelCase__ ( self , A_ ) ->List[str]: '''simple docstring''' import jax import jax.numpy as jnp if isinstance(A_ , A_ ) and column: if all( isinstance(A_ , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(A_ , axis=0 ) return column def UpperCamelCase__ ( self , A_ ) ->List[Any]: '''simple docstring''' import jax import jax.numpy as jnp if isinstance(A_ , (str, bytes, type(A_ )) ): return value elif isinstance(A_ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() __lowerCAmelCase : Tuple = {} if isinstance(A_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: __lowerCAmelCase : Dict = {'''dtype''': jnp.intaa} else: __lowerCAmelCase : List[str] = {'''dtype''': jnp.intaa} elif isinstance(A_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): __lowerCAmelCase : List[Any] = {'''dtype''': jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A_ , PIL.Image.Image ): __lowerCAmelCase : List[str] = np.asarray(A_ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: __lowerCAmelCase : Tuple = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(A_ , **{**default_dtype, **self.jnp_array_kwargs} ) def UpperCamelCase__ ( self , A_ ) ->Any: '''simple docstring''' import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(A_ , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(A_ , '''__array__''' ) and not isinstance(A_ , jax.Array ): __lowerCAmelCase : str = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A_ , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A_ ) for substruct in data_struct] ) elif isinstance(A_ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(A_ ) for substruct in data_struct] ) return self._tensorize(A_ ) def UpperCamelCase__ ( self , A_ ) ->Optional[int]: '''simple docstring''' return map_nested(self._recursive_tensorize , A_ , map_list=A_ ) def UpperCamelCase__ ( self , A_ ) ->Mapping: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.numpy_arrow_extractor().extract_row(A_ ) __lowerCAmelCase : str = self.python_features_decoder.decode_row(A_ ) return self.recursive_tensorize(A_ ) def UpperCamelCase__ ( self , A_ ) ->"jax.Array": '''simple docstring''' __lowerCAmelCase : List[Any] = self.numpy_arrow_extractor().extract_column(A_ ) __lowerCAmelCase : Optional[int] = self.python_features_decoder.decode_column(A_ , pa_table.column_names[0] ) __lowerCAmelCase : Optional[int] = self.recursive_tensorize(A_ ) __lowerCAmelCase : List[Any] = self._consolidate(A_ ) return column def UpperCamelCase__ ( self , A_ ) ->Mapping: '''simple docstring''' __lowerCAmelCase : int = self.numpy_arrow_extractor().extract_batch(A_ ) __lowerCAmelCase : Any = self.python_features_decoder.decode_batch(A_ ) __lowerCAmelCase : Optional[int] = self.recursive_tensorize(A_ ) for column_name in batch: __lowerCAmelCase : Dict = self._consolidate(batch[column_name] ) return batch
275
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _UpperCamelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _UpperCamelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : str = random.randint(0 , len(lowercase__ ) - 1 ) __lowerCAmelCase : int = parent_a[:random_slice] + parent_a[random_slice:] __lowerCAmelCase : Dict = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __lowerCAmelCase : int = random.choice(lowercase__ ) return "".join(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : str = [] # Generate more children proportionally to the fitness score. __lowerCAmelCase : str = int(parent_a[1] * 1_0_0 ) + 1 __lowerCAmelCase : Optional[Any] = 1_0 if child_n >= 1_0 else child_n for _ in range(lowercase__ ): __lowerCAmelCase : List[Any] = population_score[random.randint(0 , lowercase__ )][0] __lowerCAmelCase, __lowerCAmelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _lowercase ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __lowerCAmelCase : int = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. __lowerCAmelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __lowerCAmelCase : List[str] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowercase__ ) # Generate random starting population. __lowerCAmelCase : List[Any] = [] for _ in range(lowercase__ ): population.append(''''''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. __lowerCAmelCase, __lowerCAmelCase : Tuple = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __lowerCAmelCase : Any = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. __lowerCAmelCase : Union[str, Any] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __lowerCAmelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. __lowerCAmelCase : List[Any] = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": _UpperCamelCase = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) _UpperCamelCase = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
275
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCamelCase = { "configuration_squeezebert": [ "SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "SqueezeBertConfig", "SqueezeBertOnnxConfig", ], "tokenization_squeezebert": ["SqueezeBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["SqueezeBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ "SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "SqueezeBertForMaskedLM", "SqueezeBertForMultipleChoice", "SqueezeBertForQuestionAnswering", "SqueezeBertForSequenceClassification", "SqueezeBertForTokenClassification", "SqueezeBertModel", "SqueezeBertModule", "SqueezeBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "spiece.model"} _UpperCamelCase = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _UpperCamelCase = { "AI-Sweden/gpt-sw3-126m": 2048, "AI-Sweden/gpt-sw3-350m": 2048, "AI-Sweden/gpt-sw3-1.6b": 2048, "AI-Sweden/gpt-sw3-6.7b": 2048, "AI-Sweden/gpt-sw3-20b": 2048, } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = VOCAB_FILES_NAMES _UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase = ["""input_ids""", """attention_mask"""] def __init__( self , A_ , A_=False , A_=False , A_=False , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCAmelCase : int = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCAmelCase : Union[str, Any] = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCAmelCase : str = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCAmelCase : Any = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token __lowerCAmelCase : int = eos_token if bos_token is None else bos_token else: __lowerCAmelCase : Optional[int] = '''<pad>''' if pad_token is None else pad_token __lowerCAmelCase : List[str] = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=A_ , remove_space=A_ , keep_accents=A_ , bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __lowerCAmelCase : Union[str, Any] = do_lower_case __lowerCAmelCase : Union[str, Any] = remove_space __lowerCAmelCase : int = keep_accents __lowerCAmelCase : Union[str, Any] = vocab_file __lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # Used for whitespace normalization in input texts # fmt : off __lowerCAmelCase : List[Any] = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCAmelCase : int = re.compile( f"""[{"".join(map(A_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.__dict__.copy() __lowerCAmelCase : List[Any] = None return state def __setstate__( self , A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return len(self.sp_model ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : int = self.non_printing_characters_re.sub('''''' , A_ ) # Normalize whitespaces __lowerCAmelCase : List[str] = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCAmelCase : Tuple = unicodedata.normalize('''NFC''' , A_ ) return text def UpperCamelCase__ ( self , A_ , **A_ ) ->List[str]: '''simple docstring''' __lowerCAmelCase : int = self.preprocess_text(A_ ) return self.sp_model.encode(A_ , out_type=A_ ) def UpperCamelCase__ ( self , A_ ) ->int: '''simple docstring''' return self.sp_model.PieceToId(A_ ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.IdToPiece(A_ ) @staticmethod def UpperCamelCase__ ( A_ ) ->str: '''simple docstring''' return out_string def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = [] __lowerCAmelCase : Tuple = '''''' __lowerCAmelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A_ ) + token __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(A_ ) __lowerCAmelCase : str = False out_string += self.sp_model.decode(A_ ) return out_string def UpperCamelCase__ ( self ) ->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : str = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , A_ , A_ = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase : Any = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , '''wb''' ) as fi: __lowerCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def UpperCamelCase__ ( self , A_ , A_ = False ) ->Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(A_ , A_ ): __lowerCAmelCase : Optional[Any] = self.preprocess_text(A_ ) __lowerCAmelCase : Dict = self.sp_model.encode(A_ ) else: __lowerCAmelCase : Dict = [self.preprocess_text(A_ ) for t in text] __lowerCAmelCase : Optional[int] = self.sp_model.encode(A_ ) if return_tensors is True or return_tensors == "pt": __lowerCAmelCase : Tuple = torch.tensor(A_ ) return token_ids def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' return self.sp_model.decode(A_ ) def UpperCamelCase__ ( self , A_ ) ->List[int]: '''simple docstring''' __lowerCAmelCase : int = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCAmelCase : Any = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(A_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=A_ )
275
1
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("3.7"): raise ImportWarning( "To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition." ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( "To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n" "If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`." ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
275
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """table-transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) ->Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowerCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(A_ , A_ ): __lowerCAmelCase : int = backbone_config.get('''model_type''' ) __lowerCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __lowerCAmelCase : Any = config_class.from_dict(A_ ) # set timm attributes to None __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : List[str] = None, None, None __lowerCAmelCase : Tuple = use_timm_backbone __lowerCAmelCase : Optional[Any] = backbone_config __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Tuple = num_queries __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_ffn_dim __lowerCAmelCase : Optional[int] = encoder_layers __lowerCAmelCase : List[str] = encoder_attention_heads __lowerCAmelCase : str = decoder_ffn_dim __lowerCAmelCase : Union[str, Any] = decoder_layers __lowerCAmelCase : Any = decoder_attention_heads __lowerCAmelCase : Optional[int] = dropout __lowerCAmelCase : Any = attention_dropout __lowerCAmelCase : Tuple = activation_dropout __lowerCAmelCase : Optional[Any] = activation_function __lowerCAmelCase : List[str] = init_std __lowerCAmelCase : Tuple = init_xavier_std __lowerCAmelCase : Any = encoder_layerdrop __lowerCAmelCase : List[Any] = decoder_layerdrop __lowerCAmelCase : Optional[Any] = encoder_layers __lowerCAmelCase : Optional[Any] = auxiliary_loss __lowerCAmelCase : Optional[Any] = position_embedding_type __lowerCAmelCase : Tuple = backbone __lowerCAmelCase : Any = use_pretrained_backbone __lowerCAmelCase : int = dilation # Hungarian matcher __lowerCAmelCase : Dict = class_cost __lowerCAmelCase : List[str] = bbox_cost __lowerCAmelCase : int = giou_cost # Loss coefficients __lowerCAmelCase : Optional[Any] = mask_loss_coefficient __lowerCAmelCase : Tuple = dice_loss_coefficient __lowerCAmelCase : int = bbox_loss_coefficient __lowerCAmelCase : List[Any] = giou_loss_coefficient __lowerCAmelCase : int = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self.d_model class __lowercase (_UpperCAmelCase ): _UpperCamelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def UpperCamelCase__ ( self ) ->float: '''simple docstring''' return 1e-5 @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 12
275
1
from ...configuration_utils import PretrainedConfig _UpperCamelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """tapas""" def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=1024 , A_=[3, 256, 256, 2, 256, 256, 10] , A_=0.02 , A_=1e-12 , A_=0 , A_=10.0 , A_=0 , A_=1.0 , A_=None , A_=1.0 , A_=False , A_=None , A_=1.0 , A_=1.0 , A_=False , A_=False , A_="ratio" , A_=None , A_=None , A_=64 , A_=32 , A_=False , A_=True , A_=False , A_=False , A_=True , A_=False , A_=None , A_=None , **A_ , ) ->Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=A_ , **A_ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) __lowerCAmelCase : Tuple = vocab_size __lowerCAmelCase : Optional[int] = hidden_size __lowerCAmelCase : List[str] = num_hidden_layers __lowerCAmelCase : Optional[Any] = num_attention_heads __lowerCAmelCase : Optional[Any] = hidden_act __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : Any = hidden_dropout_prob __lowerCAmelCase : str = attention_probs_dropout_prob __lowerCAmelCase : List[Any] = max_position_embeddings __lowerCAmelCase : Optional[int] = type_vocab_sizes __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : Dict = layer_norm_eps # Fine-tuning task hyperparameters __lowerCAmelCase : Any = positive_label_weight __lowerCAmelCase : Any = num_aggregation_labels __lowerCAmelCase : Optional[int] = aggregation_loss_weight __lowerCAmelCase : Optional[Any] = use_answer_as_supervision __lowerCAmelCase : int = answer_loss_importance __lowerCAmelCase : Tuple = use_normalized_answer_loss __lowerCAmelCase : Union[str, Any] = huber_loss_delta __lowerCAmelCase : Dict = temperature __lowerCAmelCase : Tuple = aggregation_temperature __lowerCAmelCase : List[str] = use_gumbel_for_cells __lowerCAmelCase : str = use_gumbel_for_aggregation __lowerCAmelCase : Dict = average_approximation_function __lowerCAmelCase : List[str] = cell_selection_preference __lowerCAmelCase : Optional[int] = answer_loss_cutoff __lowerCAmelCase : str = max_num_rows __lowerCAmelCase : Tuple = max_num_columns __lowerCAmelCase : Optional[Any] = average_logits_per_cell __lowerCAmelCase : int = select_one_column __lowerCAmelCase : Any = allow_empty_column_selection __lowerCAmelCase : Dict = init_cell_selection_weights_to_zero __lowerCAmelCase : Tuple = reset_position_index_per_cell __lowerCAmelCase : Any = disable_per_token_loss # Aggregation hyperparameters __lowerCAmelCase : Any = aggregation_labels __lowerCAmelCase : Dict = no_aggregation_label_index if isinstance(self.aggregation_labels , A_ ): __lowerCAmelCase : Tuple = {int(A_ ): v for k, v in aggregation_labels.items()}
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class __lowercase : def __init__( self , A_ , A_=None , A_=None , A_=None , A_="resnet50" , A_=3 , A_=32 , A_=3 , A_=True , A_=True , ) ->Any: '''simple docstring''' __lowerCAmelCase : Dict = parent __lowerCAmelCase : Union[str, Any] = out_indices if out_indices is not None else [4] __lowerCAmelCase : str = stage_names __lowerCAmelCase : Union[str, Any] = out_features __lowerCAmelCase : int = backbone __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = image_size __lowerCAmelCase : Union[str, Any] = num_channels __lowerCAmelCase : Optional[Any] = use_pretrained_backbone __lowerCAmelCase : Dict = is_training def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def UpperCamelCase__ ( self , A_ , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : str = TimmBackbone(config=A_ ) model.to(A_ ) model.eval() with torch.no_grad(): __lowerCAmelCase : Any = model(A_ ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() __lowerCAmelCase, __lowerCAmelCase : Tuple = config_and_inputs __lowerCAmelCase : str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch @require_timm class __lowercase (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = (TimmBackbone,) if is_torch_available() else () _UpperCamelCase = {"""feature-extraction""": TimmBackbone} if is_torch_available() else {} _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : List[Any] = TimmBackboneModelTester(self ) __lowerCAmelCase : Union[str, Any] = ConfigTester(self , config_class=A_ , has_text_modality=A_ ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = '''resnet18''' __lowerCAmelCase : Union[str, Any] = '''microsoft/resnet-18''' __lowerCAmelCase : Any = AutoBackbone.from_pretrained(A_ , use_timm_backbone=A_ ) __lowerCAmelCase : int = AutoBackbone.from_pretrained(A_ ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) __lowerCAmelCase : int = AutoBackbone.from_pretrained(A_ , use_timm_backbone=A_ , out_indices=[1, 2, 3] ) __lowerCAmelCase : Optional[int] = AutoBackbone.from_pretrained(A_ , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip('''TimmBackbone doesn\'t support feed forward chunking''' ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t have num_hidden_layers attribute''' ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' pass @unittest.skip('''TimmBackbone initialization is managed on the timm side''' ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' pass @unittest.skip('''TimmBackbone model cannot be created without specifying a backbone checkpoint''' ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t have hidden size info in its configuration.''' ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' pass @unittest.skip('''TimmBackbone doesn\'t support output_attentions.''' ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' pass @unittest.skip('''Safetensors is not supported by timm.''' ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' pass def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : List[str] = model_class(A_ ) __lowerCAmelCase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase : Union[str, Any] = [*signature.parameters.keys()] __lowerCAmelCase : Union[str, Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A_ ) def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : List[Any] = self.has_attentions # no need to test all models as different heads yield the same functionality __lowerCAmelCase : List[Any] = self.all_model_classes[0] __lowerCAmelCase : str = model_class(A_ ) model.to(A_ ) __lowerCAmelCase : Optional[int] = self._prepare_for_class(A_ , A_ ) __lowerCAmelCase : Dict = model(**A_ ) __lowerCAmelCase : int = outputs[0][-1] # Encoder-/Decoder-only models __lowerCAmelCase : str = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: __lowerCAmelCase : Dict = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=A_ ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : Dict = model_class(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Any = model(**A_ ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None __lowerCAmelCase : Tuple = copy.deepcopy(A_ ) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : Tuple = model_class(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : Union[str, Any] = model(**A_ ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights __lowerCAmelCase : Union[str, Any] = copy.deepcopy(A_ ) __lowerCAmelCase : Optional[int] = False __lowerCAmelCase : Dict = model_class(A_ ) model.to(A_ ) model.eval() __lowerCAmelCase : str = model(**A_ )
275
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def _lowercase ( lowercase__ ): __lowerCAmelCase : List[str] = {} with open(lowercase__ , '''r''' ) as file: for line_number, line in enumerate(lowercase__ ): __lowerCAmelCase : Any = line.strip() if line: __lowerCAmelCase : Dict = line.split() __lowerCAmelCase : str = line_number __lowerCAmelCase : List[str] = words[0] __lowerCAmelCase : Any = value return result def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[Any] = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : Tuple = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : List[Any] = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : str = getattr(lowercase__ , lowercase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCAmelCase : Dict = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : str = shape_pointer.shape # let's reduce dimension __lowerCAmelCase : Any = value[0] else: __lowerCAmelCase : str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __lowerCAmelCase : List[str] = value elif weight_type == "weight_v": __lowerCAmelCase : int = value elif weight_type == "bias": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCAmelCase : Dict = getattr(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowercase__ ): __lowerCAmelCase : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCAmelCase : int = '''param''' if weight_type is not None and weight_type != "param": __lowerCAmelCase : Tuple = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCAmelCase : List[str] = '''.'''.join([key, hf_param_name] ) else: __lowerCAmelCase : Optional[int] = key __lowerCAmelCase : Union[str, Any] = value if '''lm_head''' in full_key else value[0] _UpperCamelCase = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None ): __lowerCAmelCase : Any = False for key, mapped_key in MAPPING.items(): __lowerCAmelCase : Tuple = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : List[str] = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : List[Any] = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : Any = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCAmelCase : int = '''weight''' else: __lowerCAmelCase : Any = None if hf_dict is not None: rename_dict(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) else: set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return is_used return is_used def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[Any] = fairseq_model.state_dict() __lowerCAmelCase : Tuple = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : int = True else: __lowerCAmelCase : Dict = load_wavaveca_layer(lowercase__ , lowercase__ , lowercase__ ) if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : List[str] = name.split('''.''' ) __lowerCAmelCase : Any = int(items[0] ) __lowerCAmelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=True , lowercase__=False ): if config_path is not None: __lowerCAmelCase : Union[str, Any] = WavaVecaConfig.from_pretrained(lowercase__ ) else: __lowerCAmelCase : Optional[int] = WavaVecaConfig() if is_seq_class: __lowerCAmelCase : Optional[Any] = read_txt_into_dict(lowercase__ ) __lowerCAmelCase : int = idalabel __lowerCAmelCase : Optional[int] = WavaVecaForSequenceClassification(lowercase__ ) __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) feature_extractor.save_pretrained(lowercase__ ) elif is_finetuned: if dict_path: __lowerCAmelCase : List[str] = Dictionary.load(lowercase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCAmelCase : List[Any] = target_dict.pad_index __lowerCAmelCase : List[Any] = target_dict.bos_index __lowerCAmelCase : Optional[int] = target_dict.eos_index __lowerCAmelCase : Any = len(target_dict.symbols ) __lowerCAmelCase : Union[str, Any] = os.path.join(lowercase__ , '''vocab.json''' ) if not os.path.isdir(lowercase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(lowercase__ ) ) return os.makedirs(lowercase__ , exist_ok=lowercase__ ) __lowerCAmelCase : Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : int = 1 with open(lowercase__ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = WavaVecaCTCTokenizer( lowercase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=lowercase__ , ) __lowerCAmelCase : List[str] = True if config.feat_extract_norm == '''layer''' else False __lowerCAmelCase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=lowercase__ , tokenizer=lowercase__ ) processor.save_pretrained(lowercase__ ) __lowerCAmelCase : str = WavaVecaForCTC(lowercase__ ) else: __lowerCAmelCase : Any = WavaVecaForPreTraining(lowercase__ ) if is_finetuned or is_seq_class: __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCAmelCase : Union[str, Any] = argparse.Namespace(task='''audio_pretraining''' ) __lowerCAmelCase : str = fairseq.tasks.setup_task(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowercase__ ) __lowerCAmelCase : int = model[0].eval() recursively_load_weights(lowercase__ , lowercase__ , not is_finetuned ) hf_wavavec.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
275
1
import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class __lowercase (_UpperCAmelCase ): _UpperCamelCase = (DPMSolverSDEScheduler,) _UpperCamelCase = 10 def UpperCamelCase__ ( self , **A_ ) ->int: '''simple docstring''' __lowerCAmelCase : List[str] = { '''num_train_timesteps''': 1100, '''beta_start''': 0.0_001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''noise_sampler_seed''': 0, } config.update(**A_ ) return config def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=A_ , beta_end=A_ ) def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=A_ ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : int = self.scheduler_classes[0] __lowerCAmelCase : str = self.get_scheduler_config() __lowerCAmelCase : Optional[int] = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase : Optional[int] = self.dummy_model() __lowerCAmelCase : int = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase : Optional[Any] = sample.to(A_ ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase : Tuple = scheduler.scale_model_input(A_ , A_ ) __lowerCAmelCase : Optional[Any] = model(A_ , A_ ) __lowerCAmelCase : str = scheduler.step(A_ , A_ , A_ ) __lowerCAmelCase : int = output.prev_sample __lowerCAmelCase : Optional[int] = torch.sum(torch.abs(A_ ) ) __lowerCAmelCase : Optional[Any] = torch.mean(torch.abs(A_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.47_821_044_921_875 ) < 1e-2 assert abs(result_mean.item() - 0.2_178_705_964_565_277 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_352_111_816_406 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_906_892_299_652 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Tuple = self.scheduler_classes[0] __lowerCAmelCase : Tuple = self.get_scheduler_config(prediction_type='''v_prediction''' ) __lowerCAmelCase : Optional[int] = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase : Optional[Any] = self.dummy_model() __lowerCAmelCase : List[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase : List[str] = sample.to(A_ ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase : Dict = scheduler.scale_model_input(A_ , A_ ) __lowerCAmelCase : Optional[int] = model(A_ , A_ ) __lowerCAmelCase : List[str] = scheduler.step(A_ , A_ , A_ ) __lowerCAmelCase : List[str] = output.prev_sample __lowerCAmelCase : Dict = torch.sum(torch.abs(A_ ) ) __lowerCAmelCase : List[Any] = torch.mean(torch.abs(A_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 124.77_149_200_439_453 ) < 1e-2 assert abs(result_mean.item() - 0.16_226_289_014_816_284 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 128.1_663_360_595_703 ) < 1e-2 assert abs(result_mean.item() - 0.16_688_326_001_167_297 ) < 1e-3 else: assert abs(result_sum.item() - 119.8_487_548_828_125 ) < 1e-2 assert abs(result_mean.item() - 0.1_560_530_662_536_621 ) < 1e-3 def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] __lowerCAmelCase : int = self.get_scheduler_config() __lowerCAmelCase : str = scheduler_class(**A_ ) scheduler.set_timesteps(self.num_inference_steps , device=A_ ) __lowerCAmelCase : Union[str, Any] = self.dummy_model() __lowerCAmelCase : Dict = self.dummy_sample_deter.to(A_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: __lowerCAmelCase : Optional[Any] = scheduler.scale_model_input(A_ , A_ ) __lowerCAmelCase : int = model(A_ , A_ ) __lowerCAmelCase : List[Any] = scheduler.step(A_ , A_ , A_ ) __lowerCAmelCase : str = output.prev_sample __lowerCAmelCase : Optional[Any] = torch.sum(torch.abs(A_ ) ) __lowerCAmelCase : List[Any] = torch.mean(torch.abs(A_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.46_957_397_460_938 ) < 1e-2 assert abs(result_mean.item() - 0.21_805_934_607_982_635 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_353_637_695_312 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_908_382_415_771 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Optional[int] = self.scheduler_classes[0] __lowerCAmelCase : Optional[Any] = self.get_scheduler_config() __lowerCAmelCase : Optional[int] = scheduler_class(**A_ , use_karras_sigmas=A_ ) scheduler.set_timesteps(self.num_inference_steps , device=A_ ) __lowerCAmelCase : int = self.dummy_model() __lowerCAmelCase : int = self.dummy_sample_deter.to(A_ ) * scheduler.init_noise_sigma __lowerCAmelCase : Optional[Any] = sample.to(A_ ) for t in scheduler.timesteps: __lowerCAmelCase : Tuple = scheduler.scale_model_input(A_ , A_ ) __lowerCAmelCase : int = model(A_ , A_ ) __lowerCAmelCase : Union[str, Any] = scheduler.step(A_ , A_ , A_ ) __lowerCAmelCase : str = output.prev_sample __lowerCAmelCase : List[str] = torch.sum(torch.abs(A_ ) ) __lowerCAmelCase : Optional[Any] = torch.mean(torch.abs(A_ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 176.66_974_135_742_188 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 177.63_653_564_453_125 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 else: assert abs(result_sum.item() - 170.3_135_223_388_672 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = tempfile.mkdtemp() # fmt: off __lowerCAmelCase : Any = ['''''', '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on __lowerCAmelCase : List[str] = dict(zip(A_ , range(len(A_ ) ) ) ) __lowerCAmelCase : Tuple = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] __lowerCAmelCase : Dict = {'''unk_token''': '''<unk>'''} __lowerCAmelCase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCAmelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A_ ) ) __lowerCAmelCase : str = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48_145_466, 0.4_578_275, 0.40_821_073], '''image_std''': [0.26_862_954, 0.26_130_258, 0.27_577_711], } __lowerCAmelCase : Optional[int] = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(A_ , A_ ) def UpperCamelCase__ ( self , **A_ ) ->Union[str, Any]: '''simple docstring''' return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='''!''' , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->Union[str, Any]: '''simple docstring''' return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='''!''' , **A_ ) def UpperCamelCase__ ( self , **A_ ) ->str: '''simple docstring''' return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __lowerCAmelCase : Optional[Any] = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' __lowerCAmelCase : str = self.get_tokenizer() __lowerCAmelCase : List[Any] = self.get_rust_tokenizer() __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Tuple = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_slow.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Optional[int] = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) __lowerCAmelCase : Union[str, Any] = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_fast.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Union[str, Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , A_ ) self.assertIsInstance(processor_fast.tokenizer , A_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , A_ ) self.assertIsInstance(processor_fast.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' __lowerCAmelCase : List[Any] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase : Any = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowerCAmelCase : Union[str, Any] = self.get_image_processor(do_normalize=A_ ) __lowerCAmelCase : Optional[Any] = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=A_ ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Tuple = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = image_processor(A_ , return_tensors='''np''' ) __lowerCAmelCase : Union[str, Any] = processor(images=A_ , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : int = self.get_image_processor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : Optional[int] = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Tuple = '''lower newer''' __lowerCAmelCase : List[str] = processor(text=A_ , return_tensors='''np''' ) __lowerCAmelCase : List[str] = tokenizer(A_ , return_tensors='''np''' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : List[str] = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Tuple = '''lower newer''' __lowerCAmelCase : str = self.prepare_image_inputs() __lowerCAmelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''google/owlvit-base-patch32''' __lowerCAmelCase : Any = OwlViTProcessor.from_pretrained(A_ ) __lowerCAmelCase : List[str] = ['''cat''', '''nasa badge'''] __lowerCAmelCase : str = processor(text=A_ ) __lowerCAmelCase : Optional[Any] = 16 self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = '''google/owlvit-base-patch32''' __lowerCAmelCase : Optional[Any] = OwlViTProcessor.from_pretrained(A_ ) __lowerCAmelCase : Dict = [['''cat''', '''nasa badge'''], ['''person''']] __lowerCAmelCase : List[Any] = processor(text=A_ ) __lowerCAmelCase : List[Any] = 16 __lowerCAmelCase : List[str] = len(A_ ) __lowerCAmelCase : Dict = max([len(A_ ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = '''google/owlvit-base-patch32''' __lowerCAmelCase : List[Any] = OwlViTProcessor.from_pretrained(A_ ) __lowerCAmelCase : Tuple = ['''cat''', '''nasa badge'''] __lowerCAmelCase : List[str] = processor(text=A_ ) __lowerCAmelCase : Dict = 16 __lowerCAmelCase : Optional[int] = inputs['''input_ids'''] __lowerCAmelCase : List[Any] = [ [4_9406, 2368, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9406, 6841, 1_1301, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.get_image_processor() __lowerCAmelCase : str = self.get_tokenizer() __lowerCAmelCase : Optional[Any] = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : List[str] = self.prepare_image_inputs() __lowerCAmelCase : Union[str, Any] = self.prepare_image_inputs() __lowerCAmelCase : Optional[Any] = processor(images=A_ , query_images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['''query_pixel_values''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : int = self.get_image_processor() __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : str = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) __lowerCAmelCase : Union[str, Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase : Any = processor.batch_decode(A_ ) __lowerCAmelCase : Dict = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ )
275
def _lowercase ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __lowerCAmelCase : int = str(bin(lowercase__ ) )[2:] # remove the leading "0b" __lowerCAmelCase : Any = str(bin(lowercase__ ) )[2:] __lowerCAmelCase : List[str] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('''1''' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
275
1
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): def __init__( self , A_ ) ->Dict: '''simple docstring''' super().__init__() __lowerCAmelCase : Union[str, Any] = nn.ModuleList(A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = False , A_ = True , ) ->Union[ControlNetOutput, Tuple]: '''simple docstring''' for i, (image, scale, controlnet) in enumerate(zip(A_ , A_ , self.nets ) ): __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = controlnet( A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ) # merge samples if i == 0: __lowerCAmelCase, __lowerCAmelCase : Tuple = down_samples, mid_sample else: __lowerCAmelCase : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(A_ , A_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase__ ( self , A_ , A_ = True , A_ = None , A_ = False , A_ = None , ) ->int: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = 0 __lowerCAmelCase : Optional[Any] = save_directory for controlnet in self.nets: controlnet.save_pretrained( A_ , is_main_process=A_ , save_function=A_ , safe_serialization=A_ , variant=A_ , ) idx += 1 __lowerCAmelCase : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase__ ( cls , A_ , **A_ ) ->Tuple: '''simple docstring''' __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Tuple = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... __lowerCAmelCase : List[str] = pretrained_model_path while os.path.isdir(A_ ): __lowerCAmelCase : Union[str, Any] = ControlNetModel.from_pretrained(A_ , **A_ ) controlnets.append(A_ ) idx += 1 __lowerCAmelCase : Optional[Any] = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(A_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(A_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(A_ )}. Expected at least {pretrained_model_path + "_0"}.""" ) return cls(A_ )
275
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): __lowerCAmelCase : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowerCAmelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: __lowerCAmelCase : Any = math.ceil(val / multiple ) * multiple return x __lowerCAmelCase : Dict = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = get_image_size(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = output_size # determine new height and width __lowerCAmelCase : Optional[Any] = output_height / input_height __lowerCAmelCase : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowerCAmelCase : str = scale_width else: # fit height __lowerCAmelCase : str = scale_height __lowerCAmelCase : Any = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) __lowerCAmelCase : Union[str, Any] = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = None , A_ = PILImageResampling.BILINEAR , A_ = False , A_ = 1 , A_ = True , A_ = 1 / 255 , A_ = True , A_ = None , A_ = None , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Union[str, Any] = size if size is not None else {'''height''': 384, '''width''': 384} __lowerCAmelCase : Dict = get_size_dict(A_ ) __lowerCAmelCase : Optional[Any] = do_resize __lowerCAmelCase : int = size __lowerCAmelCase : Dict = keep_aspect_ratio __lowerCAmelCase : List[Any] = ensure_multiple_of __lowerCAmelCase : Tuple = resample __lowerCAmelCase : Dict = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : List[Any] = do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , A_ , A_ , A_ = False , A_ = 1 , A_ = PILImageResampling.BICUBIC , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' __lowerCAmelCase : int = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase : Union[str, Any] = get_resize_output_image_size( A_ , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=A_ , multiple=A_ , ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ , ) ->Dict: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ = None , **A_ , ) ->np.ndarray: '''simple docstring''' return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->PIL.Image.Image: '''simple docstring''' __lowerCAmelCase : int = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Optional[int] = size if size is not None else self.size __lowerCAmelCase : Union[str, Any] = get_size_dict(A_ ) __lowerCAmelCase : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowerCAmelCase : Optional[int] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowerCAmelCase : Tuple = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : str = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(A_ ) for image in images] if do_resize: __lowerCAmelCase : Optional[Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_rescale: __lowerCAmelCase : Tuple = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: __lowerCAmelCase : str = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] __lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Dict = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A_ ) != len(A_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(A_ ): __lowerCAmelCase : Optional[int] = target_sizes.numpy() __lowerCAmelCase : List[str] = [] for idx in range(len(A_ ) ): __lowerCAmelCase : Any = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=A_ ) __lowerCAmelCase : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A_ ) else: __lowerCAmelCase : Any = logits.argmax(dim=1 ) __lowerCAmelCase : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
275
1
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class __lowercase (_UpperCAmelCase ): def __init__( self , A_ = None , A_ = None , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , **A_ , ) ->Dict: '''simple docstring''' __lowerCAmelCase : Dict = path_or_paths __lowerCAmelCase : Optional[Any] = split if split or isinstance(A_ , A_ ) else '''train''' __lowerCAmelCase : List[Any] = features __lowerCAmelCase : Any = cache_dir __lowerCAmelCase : str = keep_in_memory __lowerCAmelCase : Any = streaming __lowerCAmelCase : str = num_proc __lowerCAmelCase : Any = kwargs @abstractmethod def UpperCamelCase__ ( self ) ->Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]: '''simple docstring''' pass class __lowercase (_UpperCAmelCase ): def __init__( self , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , **A_ , ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = features __lowerCAmelCase : Optional[int] = cache_dir __lowerCAmelCase : str = keep_in_memory __lowerCAmelCase : Union[str, Any] = streaming __lowerCAmelCase : Optional[int] = num_proc __lowerCAmelCase : Tuple = kwargs @abstractmethod def UpperCamelCase__ ( self ) ->Union[Dataset, IterableDataset]: '''simple docstring''' pass
275
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCAmelCase : Dict = Vector() def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(A_ ) , '''(0,0,0,0,0,1)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(A_ ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Vector([1, 2] ) __lowerCAmelCase : Optional[int] = Vector([1, 2, 3, 4, 5] ) __lowerCAmelCase : Optional[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCAmelCase : str = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Vector([1, 2, 3] ) __lowerCAmelCase : List[str] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : List[Any] = Vector([2, -1, 4] ) # for test of dot product __lowerCAmelCase : Optional[int] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : str = Vector([1, 2, 3] ) __lowerCAmelCase : Any = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , A_ , A_ ) ) , '''(3,4,7)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Vector([1, 0, 0, 0, 0, 0] ) __lowerCAmelCase : Optional[Any] = x.copy() self.assertEqual(str(A_ ) , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[str] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(A_ ) , '''(0,1,0)''' ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(A_ , A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCAmelCase : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(A_ ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Any = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : Dict = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' __lowerCAmelCase : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCAmelCase : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def UpperCamelCase__ ( self ) ->None: '''simple docstring''' self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
275
1
def _lowercase ( lowercase__ = 1_0 , lowercase__ = 2_2 ): __lowerCAmelCase : str = range(1 , lowercase__ ) __lowerCAmelCase : List[Any] = range(1 , lowercase__ ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F"{solution(10, 22) = }")
275
def _lowercase ( lowercase__ , lowercase__ ): if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
275
1
def _lowercase ( lowercase__ = 1_0_0 ): __lowerCAmelCase : int = 0 __lowerCAmelCase : Dict = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"{solution() = }")
275
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
1
import os import jsonlines import numpy as np from tqdm import tqdm _UpperCamelCase = 2048 _UpperCamelCase = 4096 _UpperCamelCase = 42 _UpperCamelCase = os.environ.pop("PROCESS_TRAIN", "false") _UpperCamelCase = {"null": 0, "short": 1, "long": 2, "yes": 3, "no": 4} def _lowercase ( lowercase__ ): def choose_first(lowercase__ , lowercase__=False ): assert isinstance(lowercase__ , lowercase__ ) if len(lowercase__ ) == 1: __lowerCAmelCase : Tuple = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: __lowerCAmelCase : str = {k: [a[k]] for k in a} if len(a['''start_token'''] ) > 0: break return a __lowerCAmelCase : Optional[Any] = {'''id''': example['''id''']} __lowerCAmelCase : List[str] = example['''annotations'''] __lowerCAmelCase : Optional[int] = annotation['''yes_no_answer'''] if 0 in yes_no_answer or 1 in yes_no_answer: __lowerCAmelCase : int = ['''yes'''] if 1 in yes_no_answer else ['''no'''] __lowerCAmelCase : Optional[int] = [] __lowerCAmelCase : Any = [] __lowerCAmelCase : List[str] = ['''<cls>'''] else: __lowerCAmelCase : Any = ['''short'''] __lowerCAmelCase : List[str] = choose_first(annotation['''short_answers'''] ) if len(out['''start_token'''] ) == 0: # answer will be long if short is not available __lowerCAmelCase : int = ['''long'''] __lowerCAmelCase : int = choose_first(annotation['''long_answer'''] , is_long_answer=lowercase__ ) __lowerCAmelCase : Dict = [] answer.update(lowercase__ ) # disregard some samples if len(answer['''start_token'''] ) > 1 or answer["start_token"] == answer["end_token"]: __lowerCAmelCase : Optional[Any] = True else: __lowerCAmelCase : str = False __lowerCAmelCase : str = ['''start_token''', '''end_token''', '''start_byte''', '''end_byte''', '''text'''] if not all(isinstance(answer[k] , lowercase__ ) for k in cols ): raise ValueError('''Issue in ID''' , example['''id'''] ) return answer def _lowercase ( lowercase__ , lowercase__=False ): __lowerCAmelCase : Dict = _get_single_answer(lowercase__ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element __lowerCAmelCase : Optional[Any] = example['''document''']['''tokens'''] __lowerCAmelCase : Union[str, Any] = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) return { "context": " ".join(lowercase__ ), "answer": { "start_token": -1_0_0, # ignore index in cross-entropy "end_token": -1_0_0, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples __lowerCAmelCase : Any = ['''start_token''', '''end_token'''] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 __lowerCAmelCase : Optional[Any] = example['''document''']['''tokens'''] __lowerCAmelCase : List[Any] = answer['''start_token'''] __lowerCAmelCase : str = answer['''end_token'''] __lowerCAmelCase : Dict = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 __lowerCAmelCase : Optional[Any] = ''' '''.join(context[start_token:end_token] ) # checking above code if assertion: __lowerCAmelCase : List[Any] = doc['''is_html'''][answer['''start_token'''] : answer['''end_token''']] __lowerCAmelCase : List[str] = doc['''token'''][answer['''start_token'''] : answer['''end_token''']] __lowerCAmelCase : Optional[Any] = ''' '''.join([old[i] for i in range(len(lowercase__ ) ) if not is_html[i]] ) if new != old: print('''ID:''' , example['''id'''] ) print('''New:''' , lowercase__ , end='''\n''' ) print('''Old:''' , lowercase__ , end='''\n\n''' ) return { "context": " ".join(lowercase__ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def _lowercase ( lowercase__ , lowercase__ , lowercase__=2_0_4_8 , lowercase__=4_0_9_6 , lowercase__=True ): # overlap will be of doc_stride - q_len __lowerCAmelCase : int = get_context_and_ans(lowercase__ , assertion=lowercase__ ) __lowerCAmelCase : str = out['''answer'''] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } __lowerCAmelCase : List[Any] = tokenizer(example['''question''']['''text'''] , out['''context'''] ).input_ids __lowerCAmelCase : List[Any] = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Any = [] __lowerCAmelCase : Any = input_ids[:q_len] __lowerCAmelCase : str = range(lowercase__ , len(lowercase__ ) , max_length - doc_stride ) for i in doc_start_indices: __lowerCAmelCase : Optional[Any] = i + max_length - q_len __lowerCAmelCase : Dict = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['''category'''][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-1_0_0] * len(lowercase__ ), "end_token": [-1_0_0] * len(lowercase__ ), "category": category, }, } __lowerCAmelCase : Tuple = out['''context'''].split() __lowerCAmelCase : Tuple = splitted_context[answer['''end_token''']] __lowerCAmelCase : Optional[int] = len( tokenizer( ''' '''.join(splitted_context[: answer['''start_token''']] ) , add_special_tokens=lowercase__ , ).input_ids ) __lowerCAmelCase : str = len( tokenizer(''' '''.join(splitted_context[: answer['''end_token''']] ) , add_special_tokens=lowercase__ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token __lowerCAmelCase : Any = len(tokenizer(lowercase__ , add_special_tokens=lowercase__ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 __lowerCAmelCase : Dict = input_ids[answer['''start_token'''] : answer['''end_token'''] + 1] # right & left are inclusive __lowerCAmelCase : Optional[int] = answer['''start_token'''] __lowerCAmelCase : str = answer['''end_token'''] if assertion: __lowerCAmelCase : Optional[int] = tokenizer.decode(lowercase__ ) if answer["span"] != new: print('''ISSUE IN TOKENIZATION''' ) print('''OLD:''' , answer['''span'''] ) print('''NEW:''' , lowercase__ , end='''\n\n''' ) if len(lowercase__ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } __lowerCAmelCase : int = input_ids[:q_len] __lowerCAmelCase : Tuple = range(lowercase__ , len(lowercase__ ) , max_length - doc_stride ) __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Dict = [] __lowerCAmelCase : int = [] __lowerCAmelCase : Optional[Any] = [] # null, yes, no, long, short for i in doc_start_indices: __lowerCAmelCase : int = i + max_length - q_len __lowerCAmelCase : Tuple = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: __lowerCAmelCase : List[Any] = start_token - i + q_len __lowerCAmelCase : Union[str, Any] = end_token - i + q_len answers_category.append(answer['''category'''][0] ) # ["short"] -> "short" else: __lowerCAmelCase : Any = -1_0_0 __lowerCAmelCase : Tuple = -1_0_0 answers_category.append('''null''' ) __lowerCAmelCase : int = inputs[-1][start_token : end_token + 1] answers_start_token.append(lowercase__ ) answers_end_token.append(lowercase__ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('''ISSUE in strided for ID:''' , example['''id'''] ) print('''New:''' , tokenizer.decode(lowercase__ ) ) print('''Old:''' , tokenizer.decode(lowercase__ ) , end='''\n\n''' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def _lowercase ( lowercase__ , lowercase__ , lowercase__=2_0_4_8 , lowercase__=4_0_9_6 , lowercase__=False ): __lowerCAmelCase : List[Any] = get_strided_contexts_and_ans( lowercase__ , lowercase__ , doc_stride=lowercase__ , max_length=lowercase__ , assertion=lowercase__ , ) return example def _lowercase ( lowercase__ , lowercase__ ): with jsonlines.open(lowercase__ , '''a''' ) as writer: for example in tqdm(lowercase__ , total=len(lowercase__ ) , desc='''Saving samples ... ''' ): __lowerCAmelCase : Tuple = example['''labels'''] for ids, start, end, cat in zip( example['''input_ids'''] , labels['''start_token'''] , labels['''end_token'''] , labels['''category'''] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { '''input_ids''': ids, '''start_token''': start, '''end_token''': end, '''category''': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer _UpperCamelCase = load_dataset("natural_questions") _UpperCamelCase = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base") _UpperCamelCase = data["train" if PROCESS_TRAIN == "true" else "validation"] _UpperCamelCase = { "tokenizer": tokenizer, "doc_stride": DOC_STRIDE, "max_length": MAX_LENGTH, "assertion": False, } _UpperCamelCase = data.map(prepare_inputs, fn_kwargs=fn_kwargs) _UpperCamelCase = data.remove_columns(["annotations", "document", "id", "question"]) print(data) np.random.seed(SEED) _UpperCamelCase = "nq-training.jsonl" if PROCESS_TRAIN == "true" else "nq-validation.jsonl" save_to_disk(data, file_name=cache_file_name)
275
def _lowercase ( lowercase__ = 2_0_0 ): __lowerCAmelCase : Union[str, Any] = [1, 2, 5, 1_0, 2_0, 5_0, 1_0_0, 2_0_0] __lowerCAmelCase : Dict = [0] * (pence + 1) __lowerCAmelCase : Optional[int] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(lowercase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
275
1